Passivity Analysis of Markovian Jumping Neural Networks with Leakage Time-Varying Delays
Abstract
This paper is concerned with the passivity analysis of Markovian jumping neural networks with leakage time-varying delays. Based on a Lyapunov functional that accounts for the mixed time delays, a leakage delay-dependent passivity conditions are derived in terms of linear matrix inequalities (LMIs). The mixed delays includes leakage time-varying delays, discrete time-varying delays, and distributed time-varying delays. By employing a novel Lyapunov-Krasovskii functional having triple-integral terms, new passivity leakage delay-dependent criteria are established to guarantee the passivity performance. This performance not only depends on the upper bound of the time-varying leakage delay σ(t) but also depends on the upper bound of the derivative of the time-varying leakage delay σμ. While estimating the upper bound of derivative of the Lyapunov-Krasovskii functional, the discrete and distributed delays should be treated so as to appropriately develop less conservative results. Two numerical examples are given to show the validity and potential of the developed criteria.
1. Introduction
In the past few decades, neural networks (NNs) have been a hot research topic because of their emerged application in static image processing, pattern recognition, fixed-point computation, associative memory, combinatorial optimization [1–5]. Because the interactions between neurons are generally asynchronous in biological and artificial neural networks, time delays are usually encountered. Since the existence of time delays is frequently one of the main sources of instability for neural networks, the stability analysis for delayed neural networks had been extensively studied and many papers have been published on various types of neural networks with time delays based on the LMI approach [6–14].
On the other hand, the main idea of passivity theory is that the passive properties of a system can keep the system internally stable. In addition, passivity theory is frequently used in control systems to prove the stability of systems. The problem of passivity performance analysis has also been extensively applied in many areas such as signal processing, fuzzy control, sliding mode control [15], and networked control [16]. The passivity idea is a promising approach to the analysis of the stability of NNs, because it can lead to more general stability results. It is important to investigate the passivity analysis for neural networks with time delays. More recently, dissipativity or passivity performances of NNs have received increasing attention and many research results have been reported in the literature, for example, [17–21].
In practice, the RNNs often exhibit the behavior of finite state representations (also called clusters, patterns, or modes) which are referred to as the information latching problems [22]. In this case, the network states may switch (or jump) between different RNN modes according to a Markovian chain, and this gives rise to the so-called Markovian jumping recurrent neural networks. It has been shown that the information latching phenomenon is recognized to exist universally in neural networks [23, 24], which can be dealt with extracting finite state representation from a trained network, that is, a neural network sometimes has finite modes that switch from one to another at different times. The results related to all kinds of Markovian jump neural networks with time delay can also be found in [25–27] and the references therein. It should be pointed out that all the above mentioned references assume that the considered transition probabilities in the Markov process or Markov chain are time invariant, that is, the considered Markov process or Markov chain is assumed to be homogeneous. It is noted that such kind of assumption is required in most existing results on Markovian jump systems [28, 29]. The detailed discussion about piecewise homogeneous and nonhomogeneous Markovian jumping parameters has been given in [30] and references therein.
On the other hand, a typical time delay called as leakage (or “forgetting”) delay may exist in the negative feedback terms of the neural network and it has a great impact on the dynamic behaviors of delayed neural networks and more details are given in [31–36]. In [34] the authors introduced leakage time-varying delay for dynamical systems with nonlinear perturbations and derived leakage delay-dependent stability conditions via constructing a new type of Lyapunov-Krasovskii functional and LMI approach. Recently, the passivity analysis for neural networks of neutral type with Markovian jumping parameters and time delay in the leakage term have been addressed in [37]. With reference to the results above, it has been studied that many results get to be found out for passivity analysis of Markovian jumping neural networks with leakage time-varying delays. Thus, the main purpose of this paper is to shorten such a gap by making the first attempt to deal with the passivity analysis problem for a type of continuous-time neural networks with time-varying transition probabilities and mixed time delays.
In this paper, the problem of passivity analysis of Markovian jump neural networks with leakage time-varying delay and discrete and distributed time-varying delays is considered. The Markov process in the under lying neural networks is assumed to be finite piecewise homogeneous, which is a special nonhomogeneous (time-varying) Markov chain. Motivated by [30] a novel Lyapunov-Krasovskii functional is constructed in which the positive definite matrices are dependent on the system mode and a triple-integral term is introduced for deriving the delay-dependent stability conditions. By employing a novel Lyapunov-Krasovskii functional having triple integral terms, new passivity leakage delay-dependent criteria are established to guarantee the passivity performance of the given systems. This performance not only depends on the upper bound of the time-varying leakage delay σ(t) but also depends on the upper bound of the derivative of the time-varying leakage delay σμ. When estimating an upper bound of the derivative of the Lyapunov-Krasovskii functional, we handle the terms related to the discrete and distributed delays appropriately so as to develop less conservative results. Two numerical examples are given to show the validity and potential of the development of the proposed passivity criteria.
Notations. Let ℝn denote the n-dimensional Euclidean space and the superscript “T” denotes the transpose of a matrix or vector. I denote the identity matrix with compatible dimensions. For square matrices M1 and M2, the notation M1 > (≥, <, ≤) M2 denotes positive-definite (positive-semidefinite, negative, negative semidefinite) matrix. Let (Ω, 𝔉, P) be a complete probability space with a natural filtration {𝔉t} t≥0 and E[·] stand for the correspondent expectation operator with respect to the given probability measure P. Also, let τ > 0 and C([−τ, 0]; ℝn) denote the family of continuously differentiable function ϕ from [−τ, 0] to ℝn with the uniform norm ∥ϕ∥τ = max {max −τ≤θ≤0|ϕ(θ)|, max −τ≤θ≤0|ϕ′(θ)|}.
2. Problem Description and Preliminaries
In this paper, we make the following assumption, definition, and lemmas for deriving the main result.
Assumption 1. Each activation function fi(·) in (1) is continuous and bounded and satisfies
Lemma 2 (Jensen Inequality). For any matrix M ≥ 0, any scalars a and b with a ≤ b and a vector function x(t):[a, b] → ℛn such that the integrals concerned are well defined, the following inequality holds:
Lemma 3. For any constant matrix Z = ZT > 0 and scalars σ > 0, τ1 > 0, τ2 > 0 such that the following inequalities hold
The main purpose of this paper is to establish a delay-dependent sufficient condition to ensure that neural networks (1) are passive.
3. Main Results
Theorem 5. The given Markovian jumping neural networks (1) is passive if there exist
Proof. Denote ζ = [x(t) T g(x(t)) T] T and consider the following Lyapunov-Krasovskii functional for neural network (1):
Hence we can obtain from (10) that,
Using Dynkin’s formula, we have
Thus, if (33) holds, then since and V(x0, r0, η0) = 0 holds under zero initial condition, from (31) it follows that J(tp) ≤ 0 for any tp ≥ 0, which implies that (13) is satisfied and therefore the delayed neural networks (1) are locally passive. Next we shall prove that 𝔼[∥x(t)∥2] → 0 as t → ∞. Taking expectation on both sides of (28) and integrating from 0 to t we have
Remark 6. When σ(t) = σ, the system (1) becomes
Now, the passivity condition for the neural networks (43) is given in the following corollary and the result follows from Theorem 5.
Corollary 7. Neural networks (43) are passive if there exist
4. Problem without Switching
4.1. Description and Preliminaries
In this section, we derive passivity criterion for the delayed neural networks using the Lyapunov-Krasovskii functional without Markovian jumping parameters.
Theorem 8. Neural network (49) is passive if there exist
Proof. Denote ζ = [x(t) T g(x(t)) T] T and consider the following Lyapunov-Krasovskii functional for neural network (49):
Remark 9. In this paper, Theorem 5 provides passivity criteria for the Markovian jumping neural networks with leakage time varying delays. Such stability criterion is derived based on the assumption that the leakage time varying delays are differentiable and the values of σμ are known. A new set of triple integral terms have been introduced in the Lyapunov-Krasovskii functional to derive the leakage delay-dependent passivity conditions via LMI approach. New type of Lyapunov-Krasovskii functional is constructed in which the positive definite matrices Q1i,m, Q2i,m, Q3i,m are dependent on the system mode and a triple-integral term is introduced for deriving the delay-dependent passivity conditions.
5. Numerical Examples
In this chapter, we provide two simple examples presented here in order to illustrate the usefulness of our main results. Our aim is to examine the passivity analysis of given delayed neural networks.
Example 1. Consider the delayed neural networks (1) with the following parameters and having Markovian jumping parameters as below:
Example 2. Consider the delayed neural network (49) with the following parameters and without markovian jumping parameters as below:
6. Conclusion
In this paper, stochastic stability analysis of Markovian jump neural networks with leakage time-varying delay and discrete and distributed time-varying delays is considered. The Markov process in the underlying neural networks is finite piecewise homogeneous. A leakage delay-dependent passivity conditions have been derived in terms of LMIs by constructing novel Lyapunov-Krasovskii functional having triple integral terms. This performance not only depends on the upper bound of the time-varying leakage delay σ(t) but also depends on the upper bound of the derivative of the time-varying leakage delay σμ. Two numerical examples have been provided to demonstrate the effectiveness of the proposed methods for both with and without Markovian jumping parameters.