Exploration on robustness of exponentially global stability of recurrent neural networks with neutral terms and generalized piecewise constant arguments (Q2046990)

From MaRDI portal
Revision as of 21:12, 30 July 2023 by Importer (talk | contribs) (‎Created a new Item)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
scientific article
Language Label Description Also known as
English
Exploration on robustness of exponentially global stability of recurrent neural networks with neutral terms and generalized piecewise constant arguments
scientific article

    Statements

    Exploration on robustness of exponentially global stability of recurrent neural networks with neutral terms and generalized piecewise constant arguments (English)
    0 references
    0 references
    0 references
    0 references
    19 August 2021
    0 references
    Summary: With a view to the interference of piecewise constant arguments (PCAs) and neutral terms (NTs) to the original system and the significant applications in the signal transmission process, we explore the robustness of the exponentially global stability (EGS) of recurrent neural network (RNN) with PCAs and NTs (NPRNN). The following challenges arise: what the range of PCAs and the scope of NTs can NPRNN tolerate to be exponentially stable. So we derive two important indicators: maximum interval length of PCAs and the scope of neutral term (NT) compression coefficient here for NPRNN to be exponentially stable. Additionally, we theoretically proved that if the interval length of PCAs and the bound of NT compression coefficient are all lower than the given results herein, the disturbed NPRNN will still remain global exponential stability. Finally, there are two numerical examples to verify the deduced results' effectiveness here.
    0 references

    Identifiers