Projected equation methods for approximate solution of large linear systems (Q1012492)

From MaRDI portal
scientific article
Language Label Description Also known as
English
Projected equation methods for approximate solution of large linear systems
scientific article

    Statements

    Projected equation methods for approximate solution of large linear systems (English)
    0 references
    0 references
    0 references
    21 April 2009
    0 references
    The motivation of this comprehensive and very well organized study comes from recent advances in the field of dynamic programming, where large systems of equations of the form: \(x=Ax+b,\; A\in\mathbb{R}^{n\times n},\; b\in\mathbb{R}^{n}\) appear in the context of evaluation of the cost of a stationary policy in a Markovian decision problems. The authors focus on the case where \(n\) is very large, and a need to consider a low approximation of a solution within a subspace \(S=\{\Phi r|\;\Phi\in\mathbb{R}^{n\times s},\; r\in\mathbb{R}^s\}, s\ll n\) is inevitable. In general, the algorithms are extensions of recent approximate dynamic programming methods, known as temporal difference methods, which solve a projected form of Bellman's equation by using simulation-based approximations to this equation, or by using a projected value iteration method. Two types of methods are considered: (i) equation approximation methods and (ii) approximate Jacobi methods. They are developed separately and later a relation of both is widely discussed. Multi-step analogs of both are developed in the meaning of methods of approximate dynamic programming. Special methods using basis functions of the form \(A^m g, m\geq 0\), where \(g\) is some vector the components of which can be exactly computed are proposed and investigated. Stochastic iterative algorithms, based on simulation, which converge to the approximate solution and are suitable for very large-dimensional problems are proposed. A selection of Markov chains is discussed and some related examples and special cases are presented. Some new algorithms for Markovian decision problems are derived. Some extensions as well as related methods are considered as well. In particular it concerns a general approach for linear least squares problems and applications to some non-linear fixed point problems. The approximate dynamic programming methods proposed for optimal stopping problems are significantly generalized. Some applications concerning special cases and related problems are outlined. Numerical experiments are not given. The list of references is adequate related to the problem itself and giving links to other areas.
    0 references
    linear equations
    0 references
    projected equations
    0 references
    dynamic programming
    0 references
    temporal differences
    0 references
    simulation
    0 references
    value iteration
    0 references
    Jacobi method
    0 references
    Markov chains
    0 references
    algorithms
    0 references
    Bellman's equation
    0 references
    iteration method
    0 references
    Markovian decision problems
    0 references
    linear least squares problems
    0 references
    nonlinear fixed point problems
    0 references
    optimal stopping problems
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references