Error bounds and convergence analysis of feasible descent methods: A general approach

From MaRDI portal
Publication:1312756

DOI10.1007/BF02096261zbMath0793.90076OpenAlexW2093417350MaRDI QIDQ1312756

Zhi-Quan Luo, Paul Tseng

Publication date: 7 February 1994

Published in: Annals of Operations Research (Search for Journal in Brave)

Full work available at URL: https://doi.org/10.1007/bf02096261



Related Items

Some recent advances in projection-type methods for variational inequalities, Globally convergent block-coordinate techniques for unconstrained optimization, Further properties of the forward-backward envelope with applications to difference-of-convex programming, Linearly convergent away-step conditional gradient for non-strongly convex functions, Convergence results of a new monotone inertial forward-backward splitting algorithm under the local Hölder error bound condition, Some developments in general variational inequalities, New trends in general variational inequalities, Convergence rate analysis of an asynchronous space decomposition method for convex Minimization, Parallel random block-coordinate forward-backward algorithm: a unified convergence analysis, On the solution of convex QPQC problems with elliptic and other separable constraints with strong curvature, A survey on some recent developments of alternating direction method of multipliers, A family of second-order methods for convex \(\ell _1\)-regularized optimization, A stochastic extra-step quasi-Newton method for nonsmooth nonconvex optimization, On linear convergence of iterative methods for the variational inequality problem, Parallel Random Coordinate Descent Method for Composite Minimization: Convergence Analysis and Error Bounds, An inexact Newton-like conditional gradient method for constrained nonlinear systems, Augmented Lagrangian methods for convex matrix optimization problems, Error bounds in mathematical programming, Kurdyka-Łojasiewicz exponent via inf-projection, Unnamed Item, Convergence analysis of perturbed feasible descent methods, From error bounds to the complexity of first-order descent methods for convex functions, A unified approach to error bounds for structured convex optimization problems, Approximation accuracy, gradient methods, and error bound for structured convex optimization, Scaling up Bayesian variational inference using distributed computing clusters, Inexact first-order primal-dual algorithms, Quadratic growth conditions and uniqueness of optimal solution to Lasso, Linear convergence of first order methods for non-strongly convex optimization, A block symmetric Gauss-Seidel decomposition theorem for convex composite quadratic programming and its applications, Quadratic Growth Conditions for Convex Matrix Optimization Problems Associated with Spectral Functions, General inertial proximal gradient method for a class of nonconvex nonsmooth optimization problems, Linear Convergence of Proximal Gradient Algorithm with Extrapolation for a Class of Nonconvex Nonsmooth Minimization Problems, Superrelaxation and the rate of convergence in minimizing quadratic functions subject to bound constraints, Nonmonotone gradient methods for vector optimization with a portfolio optimization application, Convergence of the forward-backward algorithm: beyond the worst-case with the help of geometry, A coordinate descent margin based-twin support vector machine for classification, Nonparallel hyperplane support vector machine for binary classification problems, A smoothing proximal gradient algorithm with extrapolation for the relaxation of \({\ell_0}\) regularization problem, Error bound and isocost imply linear convergence of DCA-based algorithms to D-stationarity, Extragradient method in optimization: convergence and complexity, Separable spherical constraints and the decrease of a quadratic function in the gradient projection step, On stochastic mirror-prox algorithms for stochastic Cartesian variational inequalities: randomized block coordinate and optimal averaging schemes, A global piecewise smooth Newton method for fast large-scale model predictive control, The 2-coordinate descent method for solving double-sided simplex constrained minimization problems, Nonsmooth optimization using Taylor-like models: error bounds, convergence, and termination criteria, A quasi-Newton modified LP-Newton method, A linearly convergent stochastic recursive gradient method for convex optimization, On convergence rate of the randomized Gauss-Seidel method, On the Quadratic Convergence of the Cubic Regularization Method under a Local Error Bound Condition, On the linear convergence of the approximate proximal splitting method for non-smooth convex optimization, Alternating minimization methods for strongly convex optimization, A family of inexact SQA methods for non-smooth convex minimization with provable convergence guarantees based on the Luo-Tseng error bound property, The convergence rate analysis of the symmetric ADMM for the nonconvex separable optimization problems, Local linear convergence of an ADMM-type splitting framework for equality constrained optimization, On the convergence of the gradient projection method for convex optimal control problems with bang-bang solutions, RSG: Beating Subgradient Method without Smoothness and Strong Convexity, A coordinate gradient descent method for nonsmooth separable minimization, Incrementally updated gradient methods for constrained and regularized optimization, Iteration complexity analysis of block coordinate descent methods, A coordinate gradient descent method for \(\ell_{1}\)-regularized convex minimization, Global Lipschitzian error bounds for semidefinite complementarity problems with emphasis on NCPs, A proximal gradient descent method for the extended second-order cone linear complementarity problem, A coordinate gradient descent method for linearly constrained smooth optimization and support vector machines training, New characterizations of Hoffman constants for systems of linear constraints, An optimal algorithm and superrelaxation for minimization of a quadratic function subject to separable convex constraints with applications, Resolving learning rates adaptively by locating stochastic non-negative associated gradient projection points using line searches, Local linear convergence of the alternating direction method of multipliers for nonconvex separable optimization problems, Calculus of the exponent of Kurdyka-Łojasiewicz inequality and its applications to linear convergence of first-order methods, Error bounds for non-polyhedral convex optimization and applications to linear convergence of FDM and PGM, Rate of convergence analysis of dual-based variables decomposition methods for strongly convex problems, On the linear convergence of forward-backward splitting method. I: Convergence analysis, Over relaxed hybrid proximal extragradient algorithm and its application to several operator splitting methods, An efficient Hessian based algorithm for solving large-scale sparse group Lasso problems, A parallel line search subspace correction method for composite convex optimization, Fast global convergence of gradient methods for high-dimensional statistical recovery, Computational Methods for Solving Nonconvex Block-Separable Constrained Quadratic Problems, New analysis of linear convergence of gradient-type methods via unifying error bound conditions, Faster subgradient methods for functions with Hölderian growth, Restarting the accelerated coordinate descent method with a rough strong convexity estimate, Block-coordinate gradient descent method for linearly constrained nonsmooth separable optimization, Randomized Gradient Boosting Machine, Local convergence analysis of projection-type algorithms: unified approach, Projection onto a Polyhedron that Exploits Sparsity, A Block Successive Upper-Bound Minimization Method of Multipliers for Linearly Constrained Convex Optimization, Linear Convergence of Descent Methods for the Unconstrained Minimization of Restricted Strongly Convex Functions, Accelerated iterative hard thresholding algorithm for \(l_0\) regularized regression problem, Distributed optimization for degenerate loss functions arising from over-parameterization, Variational analysis perspective on linear convergence of some first order methods for nonsmooth convex optimization problems, Convergence rates of forward-Douglas-Rachford splitting method, Growth behavior of a class of merit functions for the nonlinear complementarity problem, Quadratic optimization with orthogonality constraint: explicit Łojasiewicz exponent and linear convergence of retraction-based line-search and stochastic variance-reduced gradient methods, QPALM: a proximal augmented Lagrangian method for nonconvex quadratic programs, Coordinate descent algorithms, Network manipulation algorithm based on inexact alternating minimization, The adventures of a simple algorithm, Avoiding bad steps in Frank-Wolfe variants, On the convergence of exact distributed generalisation and acceleration algorithm for convex optimisation, A reduced proximal-point homotopy method for large-scale non-convex BQP, On the rate of convergence of alternating minimization for non-smooth non-strongly convex optimization in Banach spaces, Perturbation techniques for convergence analysis of proximal gradient method and other first-order algorithms via variational analysis, A stochastic approximation method for convex programming with many semidefinite constraints, A Unified Analysis of Descent Sequences in Weakly Convex Optimization, Including Convergence Rates for Bundle Methods, Gap functions and global error bounds for history-dependent variational-hemivariational inequalities, Tensor Canonical Correlation Analysis With Convergence and Statistical Guarantees, Zeroth-Order Regularized Optimization (ZORO): Approximately Sparse Gradients and Adaptive Sampling, On Cluster-Aware Supervised Learning: Frameworks, Convergent Algorithms, and Applications, Semi-discrete optimal transport: hardness, regularization and numerical solution, An inexact quasi-Newton algorithm for large-scale \(\ell_1\) optimization with box constraints, Gap functions and global error bounds for differential variational-hemivariational inequalities, Convergence of Gradient-Based Block Coordinate Descent Algorithms for Nonorthogonal Joint Approximate Diagonalization of Matrices, Trimmed Statistical Estimation via Variance Reduction, Newton-MR: inexact Newton method with minimum residual sub-problem solver, Faster first-order primal-dual methods for linear programming using restarts and sharpness, Convergence of an asynchronous block-coordinate forward-backward algorithm for convex composite optimization, An easily computable upper bound on the Hoffman constant for homogeneous inequality systems, Convergence rate of the relaxed CQ algorithm under Hölderian type error bound property, Convergence Rate Analysis of a Dykstra-Type Projection Algorithm, A Proximal Alternating Direction Method of Multiplier for Linearly Constrained Nonconvex Minimization, Unnamed Item, Unnamed Item, The Analysis of Alternating Minimization Method for Double Sparsity Constrained Optimization Problem, Perturbation resilience of proximal gradient algorithm for composite objectives, Modified Goldstein--Levitin--Polyak projection method for asymmetric strongly monotone variational inequalities, Bounded perturbation resilience of projected scaled gradient methods, Error Bounds, Quadratic Growth, and Linear Convergence of Proximal Methods, Quantitative Convergence Analysis of Iterated Expansive, Set-Valued Mappings, On the Nonergodic Convergence Rate of an Inexact Augmented Lagrangian Framework for Composite Convex Programming, Convergence Rates for Deterministic and Stochastic Subgradient Methods without Lipschitz Continuity, Accelerate stochastic subgradient method by leveraging local growth condition, The Supporting Halfspace--Quadratic Programming Strategy for the Dual of the Best Approximation Problem, A Global Dual Error Bound and Its Application to the Analysis of Linearly Constrained Nonconvex Optimization, Hölderian Error Bounds and Kurdyka-Łojasiewicz Inequality for the Trust Region Subproblem, First-Order Algorithms for a Class of Fractional Optimization Problems, Proximal Gradient Methods for Machine Learning and Imaging



Cites Work