This paper is concerned with analysis problem for the global exponential stability of a class of recurrent neural networks (RNNs) with mixed discrete and distributed delays. We first prove the existence and uniqueness of the equilibrium point under mild conditions, assuming neither differentiability
Exponential stability preservation in discrete-time analogues of artificial neural networks with distributed delays
β Scribed by Sannay Mohamad
- Publisher
- Elsevier Science
- Year
- 2008
- Tongue
- English
- Weight
- 280 KB
- Volume
- 215
- Category
- Article
- ISSN
- 0377-0427
No coin nor oath required. For personal study only.
β¦ Synopsis
This paper demonstrates that there is a discrete-time analogue which does not require any restriction on the size of the time-step in order to preserve the exponential stability of an artificial neural network with distributed delays. The analysis exploits an appropriate Lyapunov sequence and a discrete-time system of Halanay inequalities, and also either a Young inequality or a geometric-arithmetic mean inequality, to derive several sufficient conditions on the network parameters for the exponential stability of the analogue. The sufficiency conditions are independent of the time-step, and they correspond to those that establish the exponential stability of the continuous-time network.
π SIMILAR VOLUMES
This paper deals with the problem of stability analysis for a class of discrete-time bidirectional associative memory (BAM) neural networks with time-varying delays. By employing the Lyapunov functional and linear matrix inequality (LMI) approach, a new sufficient conditions is proposed for the glob