Pages that link to "Item:Q2910884"
From MaRDI portal
The following pages link to Smoothing and First Order Methods: A Unified Framework (Q2910884):
Displaying 50 items.
- A dual method for minimizing a nonsmooth objective over one smooth inequality constraint (Q312667) (← links)
- New results on subgradient methods for strongly convex optimization problems with a unified analysis (Q316174) (← links)
- Gradient consistency for integral-convolution smoothing functions (Q491234) (← links)
- Adaptive smoothing algorithms for nonsmooth composite convex minimization (Q523569) (← links)
- Epi-convergence properties of smoothing by infimal convolution (Q526386) (← links)
- Dynamic smoothness parameter for fast gradient methods (Q683921) (← links)
- Helping you finding an appropriate regularization process (Q1639964) (← links)
- A simplified view of first order methods for optimization (Q1650767) (← links)
- New computational guarantees for solving convex optimization problems with first order methods, via a function growth condition measure (Q1659678) (← links)
- A new fast algorithm for constrained four-directional total variation image denoising problem (Q1666647) (← links)
- Optimal subgradient algorithms for large-scale convex optimization in simple domains (Q1689457) (← links)
- On the computational efficiency of subgradient methods: a case study with Lagrangian bounds (Q1697974) (← links)
- Smoothed \(\ell_1\)-regularization-based line search for sparse signal recovery (Q1701931) (← links)
- Acceleration of the PDHGM on partially strongly convex functions (Q1703999) (← links)
- Accelerated first-order methods for hyperbolic programming (Q1717219) (← links)
- Solving structured nonsmooth convex optimization with complexity \(\mathcal {O}(\varepsilon ^{-1/2})\) (Q1752352) (← links)
- A simple algorithm for a class of nonsmooth convex-concave saddle-point problems (Q1785640) (← links)
- Neural network based on systematically generated smoothing functions for absolute value equation (Q2008074) (← links)
- Constraint qualifications for Karush-Kuhn-Tucker conditions in multiobjective optimization (Q2025290) (← links)
- A dual reformulation and solution framework for regularized convex clustering problems (Q2029898) (← links)
- An inexact proximal augmented Lagrangian framework with arbitrary linearly convergent inner solver for composite convex optimization (Q2062324) (← links)
- Superquantiles at work: machine learning applications and efficient subgradient computation (Q2070410) (← links)
- Network manipulation algorithm based on inexact alternating minimization (Q2109010) (← links)
- An accelerated coordinate gradient descent algorithm for non-separable composite optimization (Q2139254) (← links)
- An augmented Lagrangian method with constraint generation for shape-constrained convex regression problems (Q2146447) (← links)
- An adaptive primal-dual framework for nonsmooth convex minimization (Q2220901) (← links)
- A regularized smoothing method for fully parameterized convex problems with applications to convex and nonconvex two-stage stochastic programming (Q2230935) (← links)
- Fully smoothed \(\ell_1\)-\(TV\) models: bounds for the minimizers and parameter choice (Q2251224) (← links)
- Weiszfeld's method: old and new results (Q2260646) (← links)
- Accelerated proximal incremental algorithm schemes for non-strongly convex functions (Q2297863) (← links)
- A class of exact penalty functions and penalty algorithms for nonsmooth constrained optimization problems (Q2307749) (← links)
- A variable smoothing algorithm for solving convex optimization problems (Q2343070) (← links)
- Selective bi-coordinate method for limit non-smooth resource allocation type problems (Q2416621) (← links)
- A simple convergence analysis of Bregman proximal gradient algorithm (Q2419573) (← links)
- Adversarial classification via distributional robustness with Wasserstein ambiguity (Q2693647) (← links)
- A simple nearly optimal restart scheme for speeding up first-order methods (Q2696573) (← links)
- A smoothing proximal gradient algorithm with extrapolation for the relaxation of \({\ell_0}\) regularization problem (Q2696923) (← links)
- A convex analysis approach to optimal controls with switching structure for partial differential equations (Q2808060) (← links)
- A family of subgradient-based methods for convex optimization problems in a unifying framework (Q2829570) (← links)
- MAGMA: Multilevel Accelerated Gradient Mirror Descent Algorithm for Large-Scale Convex Composite Minimization (Q3179624) (← links)
- A Randomized Coordinate Descent Method with Volume Sampling (Q3300772) (← links)
- A Preconditioner for A Primal-Dual Newton Conjugate Gradient Method for Compressed Sensing Problems (Q3454470) (← links)
- On the convergence rate of scaled gradient projection method (Q4559401) (← links)
- A Level-Set Method for Convex Optimization with a Feasible Solution Path (Q4562247) (← links)
- A Smooth Primal-Dual Optimization Framework for Nonsmooth Composite Convex Minimization (Q4600841) (← links)
- Complexity of first-order inexact Lagrangian and penalty methods for conic convex programming (Q4622887) (← links)
- FOM – a MATLAB toolbox of first-order methods for solving convex optimization problems (Q4646679) (← links)
- (Q4969074) (← links)
- (Q4998940) (← links)
- Bilevel Optimization of Regularization Hyperparameters in Machine Learning (Q5014628) (← links)