Pages that link to "Item:Q3449571"
From MaRDI portal
The following pages link to Accelerated, Parallel, and Proximal Coordinate Descent (Q3449571):
Displayed 50 items.
- On the optimal order of worst case complexity of direct search (Q276315) (← links)
- An attention algorithm for solving large scale structured \(l_0\)-norm penalty estimation problems (Q825333) (← links)
- A flexible coordinate descent method (Q1639710) (← links)
- Momentum and stochastic momentum for stochastic gradient, Newton, proximal point and subspace descent methods (Q2023684) (← links)
- An accelerated directional derivative method for smooth stochastic convex optimization (Q2029381) (← links)
- Fast and safe: accelerated gradient methods with optimality certificates and underestimate sequences (Q2044479) (← links)
- An inexact proximal augmented Lagrangian framework with arbitrary linearly convergent inner solver for composite convex optimization (Q2062324) (← links)
- Parallel random block-coordinate forward-backward algorithm: a unified convergence analysis (Q2133415) (← links)
- An accelerated coordinate gradient descent algorithm for non-separable composite optimization (Q2139254) (← links)
- Oracle complexity separation in convex optimization (Q2139268) (← links)
- Primal-dual block-proximal splitting for a class of non-convex problems (Q2218923) (← links)
- Worst-case complexity of cyclic coordinate descent: \(O(n^2)\) gap with randomized version (Q2220668) (← links)
- Markov chain block coordinate descent (Q2301127) (← links)
- Restarting the accelerated coordinate descent method with a rough strong convexity estimate (Q2301128) (← links)
- On the efficiency of a randomized mirror descent algorithm in online optimization problems (Q2354481) (← links)
- Nearly linear-time packing and covering LP solvers. Nearly linear-time packing and covering LP solvers, achieving width-independence and \(=(1/\varepsilon)\)-convergence (Q2414908) (← links)
- A block symmetric Gauss-Seidel decomposition theorem for convex composite quadratic programming and its applications (Q2414911) (← links)
- Accelerating block coordinate descent methods with identification strategies (Q2419524) (← links)
- A random block-coordinate Douglas-Rachford splitting method with low computational complexity for binary logistic regression (Q2419533) (← links)
- Accelerated randomized mirror descent algorithms for composite non-strongly convex optimization (Q2420797) (← links)
- A parallel line search subspace correction method for composite convex optimization (Q2516372) (← links)
- On the convergence analysis of asynchronous SGD for solving consistent linear systems (Q2685380) (← links)
- Variance reduction for root-finding problems (Q2689823) (← links)
- An Efficient Inexact ABCD Method for Least Squares Semidefinite Programming (Q2805705) (← links)
- Coordinate descent with arbitrary sampling I: algorithms and complexity<sup>†</sup> (Q2829565) (← links)
- Coordinate descent with arbitrary sampling II: expected separable overapproximation (Q2829566) (← links)
- Optimization in High Dimensions via Accelerated, Parallel, and Proximal Coordinate Descent (Q2832112) (← links)
- Asynchronous Stochastic Coordinate Descent: Parallelism and Convergence Properties (Q2954387) (← links)
- MAGMA: Multilevel Accelerated Gradient Mirror Descent Algorithm for Large-Scale Convex Composite Minimization (Q3179624) (← links)
- A Randomized Exchange Algorithm for Computing Optimal Approximate Designs of Experiments (Q3304857) (← links)
- Accelerating Nonnegative Matrix Factorization Algorithms Using Extrapolation (Q3379602) (← links)
- An Accelerated Randomized Proximal Coordinate Gradient Method and its Application to Regularized Empirical Risk Minimization (Q3451763) (← links)
- Randomized Iterative Methods for Linear Systems (Q3456879) (← links)
- Distributed Block Coordinate Descent for Minimizing Partially Separable Functions (Q3462314) (← links)
- (Q4558169) (← links)
- (Q4558572) (← links)
- Linear Coupling: An Ultimate Unification of Gradient and Mirror Descent (Q4638051) (← links)
- On the complexity of parallel coordinate descent (Q4638927) (← links)
- A Coordinate-Descent Primal-Dual Algorithm with Large Step Size and Possibly Nonseparable Functions (Q4646445) (← links)
- Stochastic Primal-Dual Hybrid Gradient Algorithm with Arbitrary Sampling and Imaging Applications (Q4687241) (← links)
- A remark on accelerated block coordinate descent for computing the proximity operators of a sum of convex functions (Q4967326) (← links)
- On the Complexity Analysis of the Primal Solutions for the Accelerated Randomized Dual Coordinate Ascent (Q4969070) (← links)
- Accelerated First-Order Primal-Dual Proximal Methods for Linearly Constrained Composite Convex Programming (Q4976160) (← links)
- Asynchronous variance-reduced block schemes for composite non-convex stochastic optimization: block-specific steplengths and adapted batch-sizes (Q5038180) (← links)
- A New Homotopy Proximal Variable-Metric Framework for Composite Convex Minimization (Q5076711) (← links)
- On the Convergence of Stochastic Primal-Dual Hybrid Gradient (Q5081780) (← links)
- Additive Schwarz Methods for Convex Optimization as Gradient Methods (Q5110553) (← links)
- Stochastic Reformulations of Linear Systems: Algorithms and Convergence Theory (Q5112239) (← links)
- (Q5149027) (← links)
- Two Symmetrized Coordinate Descent Methods Can Be $O(n^2)$ Times Slower Than the Randomized Version (Q5162659) (← links)