Global exponential convergence and stability of gradient-based neural network for online matrix inversion (Q734897)

From MaRDI portal
scientific article
Language Label Description Also known as
English
Global exponential convergence and stability of gradient-based neural network for online matrix inversion
scientific article

    Statements

    Global exponential convergence and stability of gradient-based neural network for online matrix inversion (English)
    0 references
    0 references
    0 references
    0 references
    0 references
    14 October 2009
    0 references
    Starting point of online matrix inversions is the equation \(AX-I=0\) and the recurrent gradient based neural network \(\dot{X}(t) = -\gamma (AX(t)-I)\) for a design parameter \(\gamma\). To prove convergence and stability of this neural network, the authors use the Lyapunov function \(E(t) = \) tr\((\tilde{X}^TA^TA\tilde{X})/2\) for \(\tilde{X}(t) = X(t) - X^*\) and the exact inverse \(X^*\). For nonsingular \(A\) this neural network is shown to be exponentially convergent for any start \(X(0)\), while for singular \(A\), the method is still globally stable. In experiments this convergence behavior is verified and in case of singular \(A\), the computed solution \(X\) is found to equal the Moore-Penrose pseudoinverse of \(A\).
    0 references
    0 references
    matrix inverse
    0 references
    online matrix inversion
    0 references
    neural network
    0 references
    Lyapunov function
    0 references
    stability
    0 references
    exponential convergence
    0 references
    numerical examples
    0 references
    Moore-Penrose pseudoinverse
    0 references

    Identifiers