The following pages link to Geovani Nunes Grapiglia (Q743776):
Displaying 27 items.
- (Q494336) (redirect page) (← links)
- On the convergence and worst-case complexity of trust-region and regularization methods for unconstrained optimization (Q494338) (← links)
- Nonlinear stepsize control algorithms: complexity bounds for first- and second-order optimality (Q504812) (← links)
- A subspace version of the Powell-Yuan trust-region algorithm for equality constrained optimization (Q743777) (← links)
- On the worst-case evaluation complexity of non-monotone line search algorithms (Q1694392) (← links)
- Improved optimization methods for image registration problems (Q1717565) (← links)
- A derivative-free trust-region algorithm for composite nonsmooth optimization (Q2013620) (← links)
- A generalized worst-case complexity analysis for non-monotone line searches (Q2028039) (← links)
- Worst-case evaluation complexity of derivative-free nonmonotone line search methods for solving nonlinear systems of equations (Q2052297) (← links)
- An adaptive trust-region method without function evaluations (Q2125064) (← links)
- A subspace version of the Wang-Yuan augmented Lagrangian-trust region method for equality constrained optimization (Q2660095) (← links)
- A subgradient method with non-monotone line search (Q2696908) (← links)
- On the worst-case complexity of nonlinear stepsize control algorithms for convex unconstrained optimization (Q2815548) (← links)
- Accelerated Regularized Newton Methods for Minimizing Composite Convex Functions (Q4646444) (← links)
- Tensor Methods for Minimizing Convex Functions with Hölder Continuous Higher-Order Derivatives (Q4971023) (← links)
- Tensor methods for finding approximate stationary points of convex functions (Q5038435) (← links)
- On the complexity of an augmented Lagrangian method for nonconvex optimization (Q5077039) (← links)
- Regularized Newton Methods for Minimizing Functions with Hölder Continuous Hessians (Q5737717) (← links)
- On inexact solution of auxiliary problems in tensor methods for convex optimization (Q5859013) (← links)
- An adaptive Riemannian gradient method without function evaluations (Q6167088) (← links)
- Adaptive Third-Order Methods for Composite Convex Optimization (Q6171322) (← links)
- Quadratic regularization methods with finite-difference gradient approximations (Q6175465) (← links)
- Worst-case evaluation complexity of a derivative-free quadratic regularization method (Q6181370) (← links)
- On the Complexity of an Augmented Lagrangian Method for Nonconvex Optimization (Q6320451) (← links)
- Universal nonmonotone line search method for nonconvex multiobjective optimization problems with convex constraints (Q6659765) (← links)
- TRFD: A derivative-free trust-region method based on finite differences for composite nonsmooth optimization (Q6748437) (← links)
- Fully Adaptive Zeroth-Order Method for Minimizing Functions with Compressible Gradients (Q6763923) (← links)