Exploration on robustness of exponentially global stability of recurrent neural networks with neutral terms and generalized piecewise constant arguments (Q2046990): Difference between revisions
From MaRDI portal
Latest revision as of 10:14, 26 July 2024
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Exploration on robustness of exponentially global stability of recurrent neural networks with neutral terms and generalized piecewise constant arguments |
scientific article |
Statements
Exploration on robustness of exponentially global stability of recurrent neural networks with neutral terms and generalized piecewise constant arguments (English)
0 references
19 August 2021
0 references
Summary: With a view to the interference of piecewise constant arguments (PCAs) and neutral terms (NTs) to the original system and the significant applications in the signal transmission process, we explore the robustness of the exponentially global stability (EGS) of recurrent neural network (RNN) with PCAs and NTs (NPRNN). The following challenges arise: what the range of PCAs and the scope of NTs can NPRNN tolerate to be exponentially stable. So we derive two important indicators: maximum interval length of PCAs and the scope of neutral term (NT) compression coefficient here for NPRNN to be exponentially stable. Additionally, we theoretically proved that if the interval length of PCAs and the bound of NT compression coefficient are all lower than the given results herein, the disturbed NPRNN will still remain global exponential stability. Finally, there are two numerical examples to verify the deduced results' effectiveness here.
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references