Pages that link to "Item:Q2205970"
From MaRDI portal
The following pages link to Newton-type methods for non-convex optimization under inexact Hessian information (Q2205970):
Displaying 30 items.
- An algorithm for the minimization of nonsmooth nonconvex functions using inexact evaluations and its worst-case complexity (Q2020598) (← links)
- A generalized worst-case complexity analysis for non-monotone line searches (Q2028039) (← links)
- Adaptive regularization for nonconvex optimization using inexact function values and randomly perturbed derivatives (Q2052165) (← links)
- On the local convergence of a stochastic semismooth Newton method for nonsmooth nonconvex optimization (Q2082285) (← links)
- Linesearch Newton-CG methods for convex optimization with noise (Q2084588) (← links)
- A stochastic extra-step quasi-Newton method for nonsmooth nonconvex optimization (Q2149551) (← links)
- Tensor Bernstein concentration inequalities with an application to sample estimators for high-order moments (Q2178629) (← links)
- Inexact restoration with subsampled trust-region methods for finite-sum minimization (Q2191786) (← links)
- Cubic regularization methods with second-order complexity guarantee based on a new subproblem reformulation (Q2676160) (← links)
- Zeroth-order nonconvex stochastic optimization: handling constraints, high dimensionality, and saddle points (Q2696568) (← links)
- The impact of noise on evaluation complexity: the deterministic trust-region case (Q2696963) (← links)
- Stochastic analysis of an adaptive cubic regularization method under inexact gradient evaluations and dynamic Hessian accuracy (Q5034938) (← links)
- Stochastic Trust-Region Methods with Trust-Region Radius Depending on Probabilistic Models (Q5079553) (← links)
- First-Order Methods for Nonconvex Quadratic Minimization (Q5113167) (← links)
- Global Convergence of Policy Gradient Methods to (Almost) Locally Optimal Policies (Q5139670) (← links)
- Convergence of Newton-MR under Inexact Hessian Information (Q5148404) (← links)
- An Inertial Newton Algorithm for Deep Learning (Q5159400) (← links)
- (Q5159403) (← links)
- (Q5214226) (← links)
- Adaptive Regularization Algorithms with Inexact Evaluations for Nonconvex Optimization (Q5244400) (← links)
- A Stochastic Semismooth Newton Method for Nonsmooth Nonconvex Optimization (Q5244401) (← links)
- Convergence Analysis of Inexact Randomized Iterative Methods (Q5856678) (← links)
- Convergence analysis of a subsampled Levenberg-Marquardt algorithm (Q6047687) (← links)
- An overview of stochastic quasi-Newton methods for large-scale machine learning (Q6097379) (← links)
- Globally Convergent Multilevel Training of Deep Residual Networks (Q6108152) (← links)
- An adaptive cubic regularization algorithm for computing H- and Z-eigenvalues of real even-order supersymmetric tensors (Q6114727) (← links)
- Newton-MR: inexact Newton method with minimum residual sub-problem solver (Q6114941) (← links)
- Faster Riemannian Newton-type optimization by subsampling and cubic regularization (Q6134379) (← links)
- Adaptive sampling quasi-Newton methods for zeroth-order stochastic optimization (Q6175706) (← links)
- Derivative-free separable quadratic modeling and cubic regularization for unconstrained optimization (Q6489314) (← links)