Pages that link to "Item:Q5266534"
From MaRDI portal
The following pages link to The Use of Quadratic Regularization with a Cubic Descent Condition for Unconstrained Optimization (Q5266534):
Displaying 29 items.
- Augmented Lagrangians with constrained subproblems and convergence to second-order stationary points (Q683332) (← links)
- On the worst-case evaluation complexity of non-monotone line search algorithms (Q1694392) (← links)
- A line-search algorithm inspired by the adaptive cubic regularization framework and complexity analysis (Q1730832) (← links)
- Regional complexity analysis of algorithms for nonconvex smooth optimization (Q2020615) (← links)
- A generalized worst-case complexity analysis for non-monotone line searches (Q2028039) (← links)
- On large-scale unconstrained optimization and arbitrary regularization (Q2070329) (← links)
- A cubic regularization of Newton's method with finite difference Hessian approximations (Q2138398) (← links)
- On the use of third-order models with fourth-order regularization for unconstrained optimization (Q2182770) (← links)
- A decoupled first/second-order steps technique for nonconvex nonlinear unconstrained optimization with improved complexity bounds (Q2288191) (← links)
- A Newton-CG algorithm with complexity guarantees for smooth unconstrained optimization (Q2297654) (← links)
- Optimality condition and complexity analysis for linearly-constrained optimization without differentiability on the boundary (Q2330649) (← links)
- Convergence and evaluation-complexity analysis of a regularized tensor-Newton method for solving nonlinear least-squares problems (Q2419539) (← links)
- A Newton-like method with mixed factorizations and cubic regularization for unconstrained minimization (Q2419564) (← links)
- Adaptive Quadratically Regularized Newton Method for Riemannian Optimization (Q3176355) (← links)
- On High-order Model Regularization for Constrained Optimization (Q4602340) (← links)
- Universal Regularization Methods: Varying the Power, the Smoothness and the Accuracy (Q4629334) (← links)
- On Regularization and Active-set Methods with Complexity for Constrained Optimization (Q4641664) (← links)
- Complexity Analysis of Second-Order Line-Search Algorithms for Smooth Nonconvex Optimization (Q4641667) (← links)
- Accelerated Regularized Newton Methods for Minimizing Composite Convex Functions (Q4646444) (← links)
- On high-order model regularization for multiobjective optimization (Q5038176) (← links)
- A note on the worst-case complexity of nonlinear stepsize control methods for convex smooth unconstrained optimization (Q5085238) (← links)
- On the Complexity of an Inexact Restoration Method for Constrained Optimization (Q5210514) (← links)
- A concise second-order complexity analysis for unconstrained optimization using high-order regularized models (Q5210739) (← links)
- Iteration and evaluation complexity for the minimization of functions whose computation is intrinsically inexact (Q5235099) (← links)
- Worst-Case Complexity of TRACE with Inexact Subproblem Solutions for Nonconvex Smooth Optimization (Q6046826) (← links)
- A Newton-CG Based Barrier Method for Finding a Second-Order Stationary Point of Nonconvex Conic Optimization with Complexity Guarantees (Q6114780) (← links)
- A nonlinear conjugate gradient method with complexity guarantees and its application to nonconvex regression (Q6114952) (← links)
- Regularized Newton Method with Global \({\boldsymbol{\mathcal{O}(1/{k}^2)}}\) Convergence (Q6116237) (← links)
- A Newton-CG Based Augmented Lagrangian Method for Finding a Second-Order Stationary Point of Nonconvex Equality Constrained Optimization with Complexity Guarantees (Q6116250) (← links)