Global exponential convergence and stability of gradient-based neural network for online matrix inversion (Q734897)
From MaRDI portal
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Global exponential convergence and stability of gradient-based neural network for online matrix inversion |
scientific article |
Statements
Global exponential convergence and stability of gradient-based neural network for online matrix inversion (English)
0 references
14 October 2009
0 references
Starting point of online matrix inversions is the equation \(AX-I=0\) and the recurrent gradient based neural network \(\dot{X}(t) = -\gamma (AX(t)-I)\) for a design parameter \(\gamma\). To prove convergence and stability of this neural network, the authors use the Lyapunov function \(E(t) = \) tr\((\tilde{X}^TA^TA\tilde{X})/2\) for \(\tilde{X}(t) = X(t) - X^*\) and the exact inverse \(X^*\). For nonsingular \(A\) this neural network is shown to be exponentially convergent for any start \(X(0)\), while for singular \(A\), the method is still globally stable. In experiments this convergence behavior is verified and in case of singular \(A\), the computed solution \(X\) is found to equal the Moore-Penrose pseudoinverse of \(A\).
0 references
matrix inverse
0 references
online matrix inversion
0 references
neural network
0 references
Lyapunov function
0 references
stability
0 references
exponential convergence
0 references
numerical examples
0 references
Moore-Penrose pseudoinverse
0 references
0 references
0 references
0 references