Gradient methods for minimizing composite functions
From MaRDI portal
(Redirected from Publication:359630)
Recommendations
Cites work
- scientific article; zbMATH DE number 729680 (Why is no real title available?)
- scientific article; zbMATH DE number 1906319 (Why is no real title available?)
- scientific article; zbMATH DE number 845714 (Why is no real title available?)
- scientific article; zbMATH DE number 3381785 (Why is no real title available?)
- A generalized proximal point algorithm for certain non-convex minimization problems
- Accelerating the cubic regularization of Newton's method on convex problems
- Atomic Decomposition by Basis Pursuit
- Introductory lectures on convex optimization. A basic course.
- Just relax: convex programming methods for identifying sparse signals in noise
- Linear Inversion of Band-Limited Reflection Seismograms
- Rounding of convex sets and efficient gradient methods for linear programming problems
- Smooth minimization of non-smooth functions
Cited in
(only showing first 100 items - show all)- Perturbation resilience of proximal gradient algorithm for composite objectives
- Penalized wavelet nonparametric univariate logistic regression for irregular spaced data
- Fast and Exact Leave-One-Out Analysis of Large-Margin Classifiers
- High-Dimensional Cost-constrained Regression Via Nonconvex Optimization
- scientific article; zbMATH DE number 6860781 (Why is no real title available?)
- Accelerated residual methods for the iterative solution of systems of equations
- Randomized block proximal damped Newton method for composite self-concordant minimization
- An alternating direction method of multipliers with the BFGS update for structured convex quadratic optimization
- scientific article; zbMATH DE number 7306860 (Why is no real title available?)
- A family of subgradient-based methods for convex optimization problems in a unifying framework
- Decentralized Gradient Descent Maximization Method for Composite Nonconvex Strongly-Concave Minimax Problems
- A relaxed parameter condition for the primal-dual hybrid gradient method for saddle-point problem
- Scattered data interpolation with nonnegative preservation using bivariate splines and its application
- Proximal Newton-type methods for minimizing composite functions
- Complexity of a quadratic penalty accelerated inexact proximal point method for solving linearly constrained nonconvex composite programs
- Proximal-proximal-gradient method
- Learnable descent algorithm for nonsmooth nonconvex image reconstruction
- Fast convergence of dynamical ADMM via time scaling of damped inertial dynamics
- Optimal subgradient methods: computational properties for large-scale linear inverse problems
- Inertial accelerated primal-dual methods for linear equality constrained convex optimization problems
- PPA-like contraction methods for convex optimization: a framework using variational inequality approach
- Accelerated gradient methods for nonconvex nonlinear and stochastic programming
- Parallel coordinate descent methods for big data optimization
- A Note on Application of Nesterov’s Method in Solving Lasso-Type Problems
- Block-wise ADMM with a relaxation factor for multiple-block convex programming
- scientific article; zbMATH DE number 653038 (Why is no real title available?)
- Inexact coordinate descent: complexity and preconditioning
- Convergence analysis of positive-indefinite proximal ADMM with a Glowinski's relaxation factor
- Alternating direction method of multipliers with variable metric indefinite proximal terms for convex optimization
- A self-calibrated direct approach to precision matrix estimation and linear discriminant analysis in high dimensions
- Accelerated proximal algorithms with a correction term for monotone inclusions
- Accelerated directional search with non-Euclidean prox-structure
- Improved convergence rates and trajectory convergence for primal-dual dynamical systems with vanishing damping
- Multi-block Bregman proximal alternating linearized minimization and its application to orthogonal nonnegative matrix factorization
- Block-simultaneous direction method of multipliers: a proximal primal-dual splitting algorithm for nonconvex problems with multiple constraints
- Optimal Algorithms for Stochastic Complementary Composite Minimization
- Parameter-free accelerated gradient descent for nonconvex minimization
- Potential Function-Based Framework for Minimizing Gradients in Convex and Min-Max Optimization
- Analysis and design of optimization algorithms via integral quadratic constraints
- Sorted concave penalized regression
- Scaled, inexact, and adaptive generalized FISTA for strongly convex optimization
- Inexact model: a framework for optimization and variational inequalities
- Linear coupling: an ultimate unification of gradient and mirror descent
- Inexact basic tensor methods for some classes of convex optimization problems
- Proximal gradient methods beyond monotony
- Conditional gradient algorithms for norm-regularized smooth convex optimization
- Self-concordant inclusions: a unified framework for path-following generalized Newton-type algorithms
- On stochastic accelerated gradient with convergence rate
- Generalized self-concordant functions: a recipe for Newton-type methods
- Stochastic multilevel composition optimization algorithms with level-independent convergence rates
- A subgradient method with constant step-size for \(\ell_1\)-composite optimization
- An inertial projection and contraction algorithm for pseudomonotone variational inequalities without Lipschitz continuity
- HOSVD-TMPE: an extrapolation method for multidimensional sequences
- Conditions for linear convergence of the gradient method for non-convex optimization
- An extrapolated iteratively reweighted \(\ell_1\) method with complexity analysis
- N-mode minimal tensor extrapolation methods
- A class of modified accelerated proximal gradient methods for nonsmooth and nonconvex minimization problems
- Inexact successive quadratic approximation for regularized optimization
- Contracting proximal methods for smooth convex optimization
- Gradient methods with the exponential relaxation
- Linesearch Newton-CG methods for convex optimization with noise
- The PPA-based numerical algorithm with the \(O(1/t)\) convergence rate for variant variational inequalities
- Generalized conditional gradient for sparse estimation
- On variance reduction for stochastic smooth convex optimization with multiplicative noise
- Forward-backward envelope for the sum of two nonconvex functions: further properties and nonmonotone linesearch algorithms
- Hierarchical sparse modeling: a choice of two group Lasso formulations
- Accelerated proximal envelopes: application to componentwise methods
- Tikhonov regularization as a nonparametric method for uncertainty quantification in aggregate data problems
- Scalable Frank-Wolfe on generalized self-concordant functions via simple steps
- Subgradient regularized multivariate convex regression at scale
- An accelerated primal-dual iterative scheme for the L 2 -TV regularized model of linear inverse problems
- Majorization-minimization generalized Krylov subspace methods for \({\ell _p}\)-\({\ell _q}\) optimization applied to image restoration
- Fast inertial dynamic algorithm with smoothing method for nonsmooth convex optimization
- A relax inexact accelerated proximal gradient method for the constrained minimization problem of maximum eigenvalue functions
- Nesterov perturbations and projection methods applied to IMRT
- On a scaled symmetric Dai-Liao-type scheme for constrained system of nonlinear equations with applications
- Robust High-Dimensional Regression with Coefficient Thresholding and Its Application to Imaging Data Analysis
- General inertial proximal gradient method for a class of nonconvex nonsmooth optimization problems
- An inexact variable metric proximal point algorithm for generic quasi-Newton acceleration
- On the convergence of the forward-backward splitting method with linesearches
- Additive Schwarz methods for convex optimization as gradient methods
- Efficient search of first-order Nash equilibria in nonconvex-concave smooth min-max problems
- Generalized conjugate gradient methods for \(\ell_1\) regularized convex quadratic programming with finite convergence
- First-order methods for convex optimization
- A level-set method for convex optimization with a feasible solution path
- Hard thresholding regularised logistic regression: theory and algorithms
- A proximal strictly contractive Peaceman-Rachford splitting method for convex programming with applications to imaging
- A proximal difference-of-convex algorithm with extrapolation
- Proximal quasi-Newton methods for regularized convex optimization with linear and accelerated sublinear convergence rates
- Pathwise coordinate optimization for sparse learning: algorithm and theory
- On convergence rates of linearized proximal algorithms for convex composite optimization with applications
- Incremental majorization-minimization optimization with application to large-scale machine learning
- A primal-dual flow for affine constrained convex optimization
- Fast convergence of generalized forward-backward algorithms for structured monotone inclusions
- Second-order orthant-based methods with enriched Hessian information for sparse \(\ell _1\)-optimization
- Complexity certifications of first-order inexact Lagrangian methods for general convex programming: application to real-time MPC
- Additive Schwarz methods for convex optimization with backtracking
- Approximating the total variation with finite differences or finite elements
- Two classes of spectral three-term derivative-free method for solving nonlinear equations with application
- scientific article; zbMATH DE number 7312307 (Why is no real title available?)
This page was built for publication: Gradient methods for minimizing composite functions
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q359630)