Modified conjugate gradient method for the solution of \(Ax=b\) (Q1281489)

From MaRDI portal
Revision as of 10:07, 31 January 2024 by Import240129110113 (talk | contribs) (Added link to MaRDI item.)
scientific article
Language Label Description Also known as
English
Modified conjugate gradient method for the solution of \(Ax=b\)
scientific article

    Statements

    Modified conjugate gradient method for the solution of \(Ax=b\) (English)
    0 references
    0 references
    0 references
    31 May 1999
    0 references
    To solve \(Ax=b\) (\(A\) is a symmetric positive definite matrix) by the classical conjugate gradient method, the solution \(x_k\) in the \(k\)th step is obtained by minimizing \(\| x-x_k\| _A\) for \(x_k\in K_k(A,b) =\text{span}\{b,Ab,\ldots,A^{k-1}b\}\) where \(\| x\| _A^2=(x,Ax)\). The error can be estimated in terms of \(\kappa(A)\), the condition number of \(A\). In this paper, it is proposed to solve the system by minimizing over the Krylov subspace \(K_k(\sqrt{A},b)\). Then the error can be bounded in terms of \(\kappa(\sqrt{A})\), which is smaller and hence gives a better rate of convergence. It is shown that, once \(b\) and \(\sqrt{A}b\) are computed, then the iterates are given by a short recurrence relation (as in the classical case) which requires only one \(A\)-multiplication per iteration step.
    0 references
    0 references
    modified conjugate gradient method
    0 references
    Krylov subspace method
    0 references
    convergence
    0 references
    stability
    0 references
    condition number
    0 references

    Identifiers