Pages that link to "Item:Q5135249"
From MaRDI portal
The following pages link to An investigation of Newton-Sketch and subsampled Newton methods (Q5135249):
Displaying 25 items.
- Sub-sampled Newton methods (Q1739039) (← links)
- Linesearch Newton-CG methods for convex optimization with noise (Q2084588) (← links)
- A stochastic extra-step quasi-Newton method for nonsmooth nonconvex optimization (Q2149551) (← links)
- Subsampled nonmonotone spectral gradient methods (Q2178981) (← links)
- Inexact restoration with subsampled trust-region methods for finite-sum minimization (Q2191786) (← links)
- Random projections of linear and semidefinite problems with linear inequalities (Q2689146) (← links)
- Discriminative Bayesian filtering lends momentum to the stochastic Newton method for minimizing log-convex functions (Q2693789) (← links)
- A robust multi-batch L-BFGS method for machine learning (Q4972551) (← links)
- Stochastic analysis of an adaptive cubic regularization method under inexact gradient evaluations and dynamic Hessian accuracy (Q5034938) (← links)
- (Q5038021) (← links)
- Quasi-Newton methods for machine learning: forget the past, just sample (Q5058389) (← links)
- A Unified Adaptive Tensor Approximation Scheme to Accelerate Composite Convex Optimization (Q5131958) (← links)
- Convergence of Newton-MR under Inexact Hessian Information (Q5148404) (← links)
- An Inertial Newton Algorithm for Deep Learning (Q5159400) (← links)
- Stochastic proximal quasi-Newton methods for non-convex composite optimization (Q5198046) (← links)
- A Stochastic Semismooth Newton Method for Nonsmooth Nonconvex Optimization (Q5244401) (← links)
- Convergence Analysis of Inexact Randomized Iterative Methods (Q5856678) (← links)
- Scalable subspace methods for derivative-free nonlinear least-squares optimization (Q6038650) (← links)
- An adaptive stochastic sequential quadratic programming with differentiable exact augmented Lagrangians (Q6038658) (← links)
- Convergence analysis of a subsampled Levenberg-Marquardt algorithm (Q6047687) (← links)
- M-IHS: an accelerated randomized preconditioning method avoiding costly matrix decompositions (Q6083984) (← links)
- An overview of stochastic quasi-Newton methods for large-scale machine learning (Q6097379) (← links)
- Newton-MR: inexact Newton method with minimum residual sub-problem solver (Q6114941) (← links)
- Global optimization using random embeddings (Q6160282) (← links)
- Hessian averaging in stochastic Newton methods achieves superlinear convergence (Q6165593) (← links)