Convergence of Newton-MR under inexact Hessian information

From MaRDI portal
Publication:5148404

DOI10.1137/19M1302211zbMATH Open1458.90473arXiv1909.06224MaRDI QIDQ5148404FDOQ5148404


Authors: Yang Liu, Fred Roosta Edit this on Wikidata


Publication date: 4 February 2021

Published in: SIAM Journal on Optimization (Search for Journal in Brave)

Abstract: Recently, there has been a surge of interest in designing variants of the classical Newton-CG in which the Hessian of a (strongly) convex function is replaced by suitable approximations. This is mainly motivated by large-scale finite-sum minimization problems that arise in many machine learning applications. Going beyond convexity, inexact Hessian information has also been recently considered in the context of algorithms such as trust-region or (adaptive) cubic regularization for general non-convex problems. Here, we do that for Newton-MR, which extends the application range of the classical Newton-CG beyond convexity to invex problems. Unlike the convergence analysis of Newton-CG, which relies on spectrum preserving Hessian approximations in the sense of L"{o}wner partial order, our work here draws from matrix perturbation theory to estimate the distance between the subspaces underlying the exact and approximate Hessian matrices. Numerical experiments demonstrate a great degree of resilience to such Hessian approximations, amounting to a highly efficient algorithm in large-scale problems.


Full work available at URL: https://arxiv.org/abs/1909.06224




Recommendations




Cites Work


Cited In (7)

Uses Software





This page was built for publication: Convergence of Newton-MR under inexact Hessian information

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5148404)