New error bounds and their applications to convergence analysis of iterative algorithms (Q1584006)

From MaRDI portal
Revision as of 21:38, 19 February 2024 by RedirectionBot (talk | contribs) (‎Removed claim: reviewed by (P1447): Item:Q590416)
scientific article
Language Label Description Also known as
English
New error bounds and their applications to convergence analysis of iterative algorithms
scientific article

    Statements

    New error bounds and their applications to convergence analysis of iterative algorithms (English)
    0 references
    0 references
    2 August 2001
    0 references
    The author presents two new error bounds for optimization problems over a convex set whose objective function \(f\) is either semianalytic or \(\gamma\)-strictlyconvex, with \(\gamma\geq 1\). Then these error bounds are applied to analyze the rate of convergence of a wide class of iterative descent algorithms for the optimization problem. The analysis shows that the function sequence \(\{f(x^k)\}\) converges at least at the sublinear rate of \(k^{-\varepsilon}\) for some positive constant \(\varepsilon\), where \(k\) is the iteration index. Moreover, the distances from the iterate sequence \(\{x^k\}\) to the set of stationary points of the optimization problem converge to zero at least sublinearly.
    0 references
    nonlinear programming
    0 references
    error bounds
    0 references
    convergence
    0 references
    iterative descent algorithms
    0 references

    Identifiers