Article ID Journal Published Year Pages File Type
6864875 Neurocomputing 2018 26 Pages PDF
Abstract
This paper is concerned with the generalized dissipativity analysis for the recurrent neural networks (RNNs) with time-varying delays. The generalized dissipativity analysis contains a few previous known results, such as the passivity, [τj−1,τj],-dissipativity, H∞ performance and j=1,…,p, performance in a unified framework. The delay interval with fixed terminals is changed into a dynamical one with adjustable delay interval based on convex combination technique (CCT), which is called adjustable delay interval method (ADIM). A novel augmented Lyapunov-Krasovskii functional (LKF) comprising triple integral terms and considering more information about neuron activation functions is constructed, in which the integral interval associated with delayed variables is not fixed. We give some sufficient conditions in terms of linear matrix inequalities (LMIs) to guarantee stability and generalized dissipativity of the considered neural networks. Finally, numerical examples are provided to demonstrate the effectiveness and less conservative of the obtained theoretical results.
Related Topics
Physical Sciences and Engineering Computer Science Artificial Intelligence
Authors
, , , , , ,