Next Article in Journal
Gradient Iterative Method with Optimal Convergent Factor for Solving a Generalized Sylvester Matrix Equation with Applications to Diffusion Equations
Next Article in Special Issue
Analytical Solution of Nonlinear Fractional Volterra Population Growth Model Using the Modified Residual Power Series Method
Previous Article in Journal
Error Estimation of the Homotopy Perturbation Method to Solve Second Kind Volterra Integral Equations with Piecewise Smooth Kernels: Application of the CADNA Library
Previous Article in Special Issue
Conditions to Guarantee the Existence of the Solution to Stochastic Differential Equations of Neutral Type
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Optimization Techniques for the Construction of an Exponential Estimate for Delayed Recurrent Neural Networks

by
Vasyl Martsenyuk
*,†,‡,
Stanislaw Rajba
and
Mikolaj Karpinski
Department of Computer Science and Automatics, Faculty of Mechanical Engineering and Computer Science, University of Bielsko-Biala, 43-309 Bielsko-Biala, Poland
*
Author to whom correspondence should be addressed.
Current address: 2 Willowa, 43-309 Bielsko-Biala, Poland.
These authors contributed equally to this work.
Symmetry 2020, 12(10), 1731; https://doi.org/10.3390/sym12101731
Submission received: 31 August 2020 / Revised: 23 September 2020 / Accepted: 15 October 2020 / Published: 20 October 2020
(This article belongs to the Special Issue Ordinary and Partial Differential Equations: Theory and Applications)

Abstract

:
This work is devoted to the modeling and investigation of the architecture design for the delayed recurrent neural network, based on the delayed differential equations. The usage of discrete and distributed delays makes it possible to model the calculation of the next states using internal memory, which corresponds to the artificial recurrent neural network architecture used in the field of deep learning. The problem of exponential stability of the models of recurrent neural networks with multiple discrete and distributed delays is considered. For this purpose, the direct method of stability research and the gradient descent method is used. The methods are used consequentially. Firstly we use the direct method in order to construct stability conditions (resulting in an exponential estimate), which include the tuple of positive definite matrices. Then we apply the optimization technique for these stability conditions (or of exponential estimate) with the help of a generalized gradient method with respect to this tuple of matrices. The exponential estimates are constructed on the basis of the Lyapunov–Krasovskii functional. An optimization method of improving estimates is offered, which is based on the notion of the generalized gradient of the convex function of the tuple of positive definite matrices. The search for the optimal exponential estimate is reduced to finding the saddle point of the Lagrange function.

1. Introduction

Breakthrough results in the field of deep machine learning are obtained nowadays using recurrent neural networks (RNN). In particular, the construction of machine learning models for problems of image recognition with captioning, natural language processing and translation, was made possible by recurrent neural networks with Long Short-Term Memory (LSTM) and Gated Recurrent Units (GRU). The paper [1] offers a description of such models using ordinary differential equations. Further research has to be related to the systems with time delays as they are modeling the memory within the network units. In [2] drawing from concepts in signal processing, they formally derived the canonical RNN formulation from differential equations.
Here our study of the RNN model is based on the system with multiple discrete and distributed time-varying delays
x ˙ ( t ) = A x ( t ) + k = 1 r 1 W 1 , k g ( x ( t h k ( t ) ) ) + m = 1 r 2 W 2 , m t τ m ( t ) t g ( x ( θ ) ) d θ ,
where x ( t ) R n is the state vector, A = d i a g ( a 1 , a 2 , , a n ) is a diagonal matrix with positive entries a i > 0 . For the i-th neuron 1 / a i can be interpreted as the activity decay constant (or time constant). W 1 , k = ( w i j 1 , k ) n × n , k = 1 , r 1 ¯ , W 2 , m = ( w i j 2 , m ) n × n , m = 1 , r 2 ¯ are the synaptic connection weight matrices. The entries of W 1 , k and W 2 , m may be positive (excitatory synapses) or negative (inhibitory synapses). g ( x ( t ) ) = [ g 1 ( x ( t ) ) , g 2 ( x ( t ) ) , , g n ( x ( t ) ) ] R n is the non-decreasing activation function, which belongs to sector non-linear function class defined by
g j ( 0 ) = 0 and 0 g j ( ξ 1 ) g j ( ξ 2 ) ξ 1 ξ 2 l j , l j > 0 ,
ξ 1 , ξ 2 R , ξ 1 ξ 2 , j 1 , n ¯ and x = 0 is a fixed point of Equation (1). We let L = d i a g ( l 1 , l 2 , , l n ) is a diagonal matrix with positive entries l j > 0 .
The system (1) includes discrete and distributed time-varying delays, which are described with the help of the second and the third terms correspondingly.
The bounded differentiable functions h k ( t ) represent discrete delays of system with
0 h k ( t ) h M , k ,
and
h ˙ k ( t ) h D , k < 1 ,
k = 1 , r 1 ¯ , t > 0 . Delays h k ( t ) and τ m ( t ) have physical meaning as “controllable memory” if previous states of neurons effects on output only during some time intervals. h M , k and h D , k are bounds of the delay and its derivative for discrete delays.
The bounded functions τ m ( t ) represent distributed delays of system with 0 τ m ( t ) τ M , m , m = 1 , r 2 ¯ .
The bounded functions h k ( t ) and τ m ( t ) represent axonal signal transmission delays. The condition (3) for derivative h ˙ k ( t ) will be applied when estimating the upper right derivative of Lyapunov–Krasovskii functional (see, for example, [3]).
The initial conditions associated with system (1) are assumed to be
x ( s ) = ϕ ( s ) , s [ τ M , 0 ] , τ M : = max h M , k , k = 1 , r 1 ¯ , τ M , m , m = 1 , r 2 ¯ ,
where ϕ ( s ) C [ τ M , 0 ] .
Given any ϕ ( s ) C [ τ M , 0 ] , under the assumption (2), there exists a unique trajectory of (1) starting from ϕ [4].
Here we use the Hopfield neural network model, which includes a diagonal matrix A with positive entries, that shows the self-connection of the neuron. That is the next state of the neuron is dependent on its current state and outputs of eventually all neurons. Such a diagonal matrix is traditionally applied in stability research of continuous-time RNNs [5]. On the other hand, if we used arbitrary matrix A, we would assume the next state of a neuron is dependent on its current state as well as the states of all other neurons, which means that the internal states of all neurons are accessible from outside. It contradicts that, for example, in the case of the LSTM unit only the hidden state vector (also known as the output vector) is seen.
Following the work [2], we may give the interpretation of the model (1) from the viewpoint of signal processing, leading us to “canonical” and “non-canonical” RNNs. Namely, we have x ( t ) , the state signal vector; g ( x ( t ) ) , the readout signal vector, which is a warped version of the state signal vector; the bias parameters are omitted without loss of generality since they can be used in the transformation resulting in the homogeneous system (1). Initial state ϕ ( s ) , s [ τ M , 0 ] is considered as an input signal, thus, modeling one-to-many RNN architecture. In more general many-t0-many case, input signal vector u ( t ) , t > 0 can be used during the work of the RNN as an “input sequence”.
Although RNNs can actually be described using difference equations, it makes sense to consider continuous-time equations that describe the operation of RNNs. This is due to the fact that differential equations make it possible to better describe and understand the dynamic processes that occur. In addition, with the help of differential equations it is possible to explicitly obtain the conditions for stabilization of recurrent neural networks. This is of great importance in the design of recurrent neural networks. Ref. [5] provide a comprehensive review of the research on the continuous-time recurrent neural networks focusing on the stability of Hopfield and Cohen–Grossberg neural networks.
Note also that the corresponding recurrent neural networks can be obtained by discretizing the models based on differential equations. Thus, work [2] shows the way to construct a RNN of the LSTM type starting from the corresponding model based on differential equations with delay and further through the discretization of the so-called canonical RNN.
Recurrent neural network models have been considered from the 1980s, after the pioneering work of Hopfield [6] modeling each neuron as a linear circuit consisting of a resistor and a capacitor. Two approaches can be differed when investigating the models of recurrent neural networks in the class of delayed differential equations. The first way means research of local stability with the help of comparison with the linearised system [7,8,9,10]. The conditions of the Hopf bifurcation were obtained in [10,11]. The second approach (which is called the direct Lyapunov’s method) uses Lyapunov–Krasovkii functionals [3]. It allows us to get stability conditions constructively, which are formulated in the form of linear-matrix inequalities (LMIs). These stability conditions can be improved with the help of optimization of parameters of Lyapunov–Krasovskii functionals.
Exponential estimates of the solutions are very important when investigating the models of RNNs because they show the rate of convergence of calculations when recognizing input data. In the previous works [12,13], indirect method was developed allowing us to get exponential estimates in some general cases of the RNN models. It results in the numerical solution of quasipolynomial equation. It gives us a clear value of exponential decay, which, unfortunately, does not admit optimization and so, cannot be improved. In order to overcome this shortcoming, here we develop an optimization technique, which is based on the direct method for Liapunov–Krasovskii functionals of the special kind.

2. Exponential Estimate

Let Ω n R n × n be a set of all symmetric positive definite matrices. It is an open convex cone because:
(a)
convexity—for any P 1 Ω n , P 2 Ω n , x R n , and ξ [ 0 , 1 ] we have x ( ξ P 1 + ( 1 ξ ) P 2 ) x = ξ x P 1 x + ( 1 ξ ) x P 2 x > 0 ;
(b)
cone—for any P Ω n , x R n , and η > 0 we have η x P x > 0 .
Ω ¯ n 1 is the part of the cone Ω n contained inside the unit sphere, i.e., Ω ¯ n 1 : = { P Ω n : P 1 } . Here P is Frobenius norm of the matrix P Ω n .
Lemma 1.
Reference [14] For any constant matrix U Ω n , scalar β > 0 , vector function u : [ 0 , β ] R n such that the integrations concerned are well-defined, then
0 β u ( s ) d s U 0 β u ( s ) d s β 0 β u ( s ) U u ( s ) d s .
Lemma 2.
Reference [15] Given any real matrices W 1 , W 2 , W 3 with appropriate dimensions and a scalar β > 0 , W 3 Ω n , then the following inequality holds
W 1 W 2 + W 2 W 1 β W 1 W 3 W 1 + β 1 W 2 W 3 1 W 2 .
In the following definitions, we assume that the trivial solution of (1) be the unique equilibrium point of the model (1).
Definition 1.
The trivial solution of (1) is globally asymptotically stable if for every solution x ( t ) to the initial value problem (1)–(4), we have x ( t ) 0 as t .
Definition 2.
If there exist constants α > 0 , K > 0 , and T > 0 such that every solution x ( t ) to the initial value problem (1)–(4) always satisfies x ( t ) K e α t for all t > T , then the trivial solution of (1) is said to be globally exponentially stable.
Our research is based on the following Lyapunov–Krasovskii functional, which is an extension of the one offered in the work [3] for the case of multiple delays
V [ x t ( · ) ] = e 2 α t x ( t ) P x ( t ) + k = 1 r 1 t h k ( t ) t e 2 α s g ( x ( s ) ) Q k g ( x ( s ) ) d s + m = 1 r 2 τ M , m τ M , m 0 t + θ t e 2 α s g ( x ( s ) ) S m g ( x ( s ) ) d s d θ ,
where unknown constant α > 0 and matrices P, Q k , k = 1 , r 1 ¯ , S m , m = 1 , r 2 ¯ belong to Ω n . Here we use traditional denotation of the element of the solution of (1) as the vector-interval x t ( · ) : = { x ( t + θ ) | θ [ τ M , 0 ] } C [ τ M , 0 ] .
Theorem 1.
We assume that system (1) satisfies the following condition.
H1. Let there exist constant α > 0 and matrices P, Q k , k = 1 , r 1 ¯ , S m , m = 1 , r 2 ¯ , which belong to relint ( Ω ¯ n 1 ) , such that the symmetric matrix
Γ : = Γ 11 Γ 12 Γ 21 Γ 22 , Γ 11 : = 2 α P + A P + P A L k = 1 r 1 Q k + τ M , m 2 m = 1 r 2 S m L Ω n , Γ 12 : = e α h M , 1 1 h D , 1 P W 1 , 1 e α h M , r 1 1 h D , r 1 P W 1 , r 1 e α τ M , 1 P W 2 , 1 e α τ M , r 2 P W 2 , r 2 R n × n ( r 1 + r 2 ) , Γ 21 : = Γ 12 R n ( r 1 + r 2 ) × n , Γ 22 : = Q 1 Θ Q r 1 S 1 Θ S r 2 Ω n ( r 1 + r 2 ) , Θ R n × n is matrix of zeroes ,
belong to Ω n ( 1 + r 1 + r 2 ) .
Then the trivial solution of (1) is globally asymptotically stable.
Proof. 
Estimating right upper derivative of the functional V [ x t ( · ) ] along the solution of the system (1), we get
d V + [ x t ( · ) ] d t e 2 α t { x ( t ) [ 2 α P A P P A ] x ( t ) + k = 1 r 1 g ( x ( t h k ( t ) ) ) W 1 , k P x ( t ) + x ( t ) P k = 1 r 1 W 1 , k g ( x ( t h k ( t ) ) ) + m = 1 r 2 t τ m ( t ) t g ( x ( θ ) d θ ) W 2 , m P x ( t ) + x ( t ) P m = 1 r 2 W 2 , m t τ m ( t ) t g ( x ( θ ) ) d θ } + k = 1 r 1 e 2 α t { g ( x ( t ) ) Q k g ( x ( t ) ) e 2 α h k ( t ) g ( x ( t h k ( t ) ) ) Q k g ( x ( t h k ( t ) ) ) ( 1 h D , k ) } + m = 1 r 2 τ M , m { τ M , m e 2 α t g ( x ( t ) ) S m g ( x ( t ) ) e 2 α ( t τ M , m ) t τ m ( t ) t g ( x ( s ) ) S m g ( x ( s ) ) d s } .
Applying Lemmas 1 and 2 for estimating counterparts of (7), we have
k = 1 r 1 { g ( x ( t h k ( t ) ) ) W 1 , k P x ( t ) + x ( t ) P W 1 , k g ( x ( t h k ( t ) ) ) } = k = 1 r 1 [ e α h M , k ( 1 h D , k ) 1 / 2 g ( x ( t h k ( t ) ) ) ] [ e α h M , k ( 1 h D , k ) 1 / 2 W 1 , k P x ( t ) ] + [ e α h M , k ( 1 h D , k ) 1 / 2 x ( t ) P W 1 , k ] [ e α h M , k ( 1 h D , k ) 1 / 2 g ( x ( t h k ( t ) ) ) ] k = 1 r 1 { e 2 α h M , k ( 1 h D , k ) 1 x ( t ) P W 1 , k Q k 1 W 1 , k P x ( t ) + e 2 α h M , k ( 1 h D , k ) g ( x ( t h k ( t ) ) ) Q k g ( x ( t h k ( t ) ) ) } ,
and
m = 1 r 2 { t τ m ( t ) t g ( x ( θ ) ) d θ W 2 , m P x ( t ) + x ( t ) P W 2 , m t τ m ( t ) t g ( x ( θ ) ) d θ } = m = 1 r 2 { [ e α τ M , m t τ m ( t ) t g ( x ( θ ) ) d θ ] [ e α τ M , m W 2 , m P x ( t ) ] + [ e α τ M , m x ( t ) P W 2 , m ] [ t τ m ( t ) t g ( x ( θ ) ) d θ e α τ M , m ] } m = 1 r 2 { e 2 α τ M , m t τ m ( t ) t g ( x ( θ ) ) d θ S m t τ m ( t ) t g ( x ( θ ) ) d θ + e 2 α τ M , m x ( t ) P W 2 , m S m 1 W 2 , m P x ( t ) } m = 1 r 2 { τ M , m e 2 α τ M , m t τ m ( t ) t g ( x ( θ ) ) S m g ( x ( θ ) ) d θ + e 2 α τ M , m x ( t ) P W 2 , m S m 1 W 2 , m P x ( t ) } .
Finally we get
d V + [ x t ( · ) ] d t e 2 α t x ( t ) { 2 α P A P P A + k = 1 r 1 e 2 α h M , k ( 1 h D , k ) 1 P W 1 , k Q k 1 W 1 , k P + L Q k L + m = 1 r 2 e 2 α τ M , m P W 2 , m S m 1 W 2 , m P + τ M , m 2 L S m L } x ( t ) e 2 α t x ( t ) { Γ 11 Γ 12 Γ 22 1 Γ 21 } x ( t ) = e 2 α t x ( t ) Γ / Γ 22 x ( t ) ,
where Γ / Γ 22 : = Γ 11 Γ 12 Γ 22 1 Γ 21 is the Schur complement of Γ in Γ 22 . From the Schur complement it follows that the right side of (8) is negative definite if and only if Γ Ω n ( 1 + r 1 + r 2 ) [16]. ☐
Corollary 1.
Provided that the condition H1 holds the trivial solution of (1) is globally exponentially stable as follows
x ( t ) γ ( α ) | ϕ | τ M e α t , t > 0 ,
where
γ ( α ) : = λ min 1 / 2 ( P ) ( λ max ( P ) + k = 1 r 1 λ max ( Q k ) l max 2 1 e 2 α h M , k 2 α + m = 1 r 2 τ M , m λ max ( L S m L ) 2 α τ M , m 1 + e 2 α τ M , m 4 α 2 ) 1 / 2 , l max : = max { l 1 , , l n } ,
λ min ( · ) and λ max ( · ) are minimal and maximal eigenvalues of the matrix. Here we use denotions of · as Euclidean norm in R n and | x ( · ) | τ M : = sup s [ τ M , 0 ] x ( s ) as the uniform convergence norm in C [ τ M , 0 ] .
Proof. 
Firstly note that the inequality
2 α τ M , m + e 2 α τ M , m 1 ,
enables us that the square root expression in γ ( α ) is nonnegative for α > 0 . From Theorem 1 it follows that V [ x t ( · ) ] V [ ϕ ( · ) ] . Hence, we get
λ min ( P ) x ( t ) 2 e 2 α t V [ x t ( · ) ] e 2 α t V [ ϕ ( · ) ] e 2 α t ( ϕ ( 0 ) P ϕ ( 0 ) + k = 1 r 1 h M , k 0 e 2 α s g ( x ( s ) ) Q k g ( x ( s ) ) d s + m = 1 r 2 τ M , m τ M , m 0 θ 0 e 2 α s g ( ϕ ( s ) ) S m g ( ϕ ( s ) ) d s d θ ) e 2 α t ( λ max ( P ) + k = 1 r 1 λ max ( Q k ) l max 2 h M , k 0 e 2 α s d s + m = 1 r 2 τ M , m λ max ( L S m L ) τ M , m 0 θ 0 e 2 α s d s d θ ) | ϕ | τ M 2 = e 2 α t ( λ max ( P ) + k = 1 r 1 λ max ( Q k ) l max 2 1 e 2 α h M , k 2 α + m = 1 r 2 τ M , m λ max ( L S m L ) 2 α τ M , m 1 + e 2 α τ M , m 4 α 2 ) | ϕ | τ M 2 .
Finally, it yields
λ min ( P ) x ( t ) 2 e 2 α t γ ( α ) | ϕ | τ M 2 .

3. Optimization Method

The condition H1 describes the main stability result. The matrix Γ presents operator, which is linear with the respect to the tuple of matrices P , Q 1 , , Q r 1 , S 1 , , S r 2 . At the same time, the dependence of Γ on α is nonlinear. α is the parameter determining the exponential decay rate. Since optimization of Γ with the respect to P , Q 1 , , Q r 1 , S 1 , , S r 2 may be considered as linear matrix inequality, we can reduce the problem of the construction of exponential estimate like (9) to the convex programming. It is natural to assume that “the more positive definite” is the matrix Γ , the “more asymptotically stable” is the trivial solution. In turn, the solution is “more exponentially” stable. The positive definiteness of the matrix Γ is described with the help of minimum eigenvalue. Thus we result in the following optimization problem.
Further we apply optimization technique developed earlier for linear systems in [17]. Given α > 0 we search the tuple of matrices ( P , Q 1 , , Q r 1 , S 1 , , S r 2 ) as a solution of the optimization problem
P , Q 1 , , Q r 1 , S 1 , , S r 2 = arg inf [ P , Q 1 , , Q r 1 , S 1 , , S r 2 ] i = 1 1 + r 1 + r 2 Ω ¯ n 1 ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 .
Here ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 = λ min ( Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 ) .
We give some general conditions of existence of a solution of problem (10).
Definition 3.
The inner product of the tuples of matrices P 1 , Q 1 , 1 , , Q r 1 , 1 , S 1 , 1 , , S r 2 , 1 , P 2 , Q 1 , 2 , , Q r 1 , 2 , S 1 , 2 , , S r 2 , 2 i = 1 1 + r 1 + r 2 Ω ¯ n 1 is
P 1 , Q 1 , 1 , , Q r 1 , 1 , S 1 , 1 , , S r 2 , 1 , P 2 , Q 1 , 2 , , Q r 1 , 2 , S 1 , 2 , , S r 2 , 2 : = i , j = 1 n p i j 1 p i j 2 + k = 1 r 1 q i j , k 1 q i j , k 2 + m = 1 r 2 s i j , m 1 s i j , m 2 ,
where P δ = { p i j δ } , Q k , δ = { q i j , k δ } , S m , δ = { s i j , m δ } , i , j = 1 , n ¯ , k = 1 , r 1 ¯ , m = 1 , r 2 ¯ , δ = 1 , 2 .
Definition 4.
The generalized gradient of the convex function ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 at the interior point P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 i = 1 1 + r 1 + r 2 Ω ¯ n 1 is the tuple of matrices D 0 , E 1 , 0 , , E r 1 , 0 , F 1 , 0 , , F r 2 , 0 i = 1 1 + r 1 + r 2 R n × n such that for all ( P , Q 1 , , Q r 1 , S 1 , , S r 2 ) i = 1 1 + r 1 + r 2 Ω ¯ n 1 we have
ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 ψ 0 P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 D 0 , E 1 , 0 , , E r 1 , 0 , F 1 , 0 , , F r 2 , 0 ) , ( P P 0 , Q 1 Q 1 , 0 , , Q r 1 Q r 1 , 0 , S 1 S 1 , 0 , , S r 2 S r 2 , 0 .
Let Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 be a linear matrix-valued operator that maps the tuple of matrices P , Q 1 , , Q r 1 , S 1 , , S r 2 i = 1 1 + r 1 + r 2 Ω ¯ n 1 into the n ( 1 + r 1 + r 2 ) × n ( 1 + r 1 + r 2 ) symmetric matrices Γ .
Denote by Δ i j R n × n the matrix in which the entries at positions ( i , j ) and ( j , i ) are units, and all the rest are zeroes.
Lemma 3.
The generalized gradient of the function ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 = λ min ( Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 ) at the interior point P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 i = 1 1 + r 1 + r 2 Ω ¯ n 1 is the tuple of matrices D 0 , E 1 , 0 , , E r 1 , 0 , F 1 , 0 , , F r 2 , 0 , where D 0 = { d i j 0 } , E k , 0 = { e i j , k 0 } , F m , 0 = { f i j , m 0 } , i , j = 1 , n ¯ , k = 1 , r 1 ¯ , m = 1 , r 2 ¯ such that
d i j 0 = z 0 Γ ( Δ i j , Θ 1 , , Θ r 1 + r 2 ) z 0 , i f i = j , 1 2 z 0 Γ ( Δ i j , Θ 1 , , Θ r 1 + r 2 ) z 0 , i f i j , e i j , k 0 = z 0 Γ ( Θ , Θ 1 , , Θ k 1 , Δ i j , Θ k + 1 , , Θ r 1 , Θ r 1 + 1 , , Θ r 1 + r 2 ) z 0 , i f i = j , 1 2 z 0 Γ ( Θ , Θ 1 , , Θ k 1 , Δ i j , Θ k + 1 , , Θ r 1 , Θ r 1 + 1 , , Θ r 1 + r 2 ) z 0 , i f i j , f i j , m 0 = z 0 Γ ( Θ , Θ 1 , , Θ r 1 , Θ r 1 + 1 , , Θ m 1 , Δ i j , Θ m + 1 , , Θ r 1 + r 2 ) z 0 , i f i = j , 1 2 z 0 Γ ( Θ , Θ 1 , , Θ r 1 , Θ r 1 + 1 , , Θ m 1 , Δ i j , Θ m + 1 , , Θ r 1 + r 2 ) z 0 , i f i j ,
where Θ v is matrix Θ at position v,
z 0 is the unit eigenvector corresponding to λ min ( Γ ( P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 ) ) .
Proof. 
Consider for z R n ( 1 + r 1 + r 2 )
ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 ψ 0 P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 = min z = 1 { z Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 z } + min z = 1 { z Γ P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 z } .
Assume the first quadratic form reaches its minimal value at z = z 1 , and the second one at z = z 0 . Then subtracting and adding the expression z 0 Γ ( P , Q 1 , , Q r 1 , S 1 , , S r 2 ) z 0 , we get
ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 ψ 0 P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 = z 0 Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 Γ P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 z 0 + z 0 Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 z 0 z 1 Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 z 1 .
Since z 0 Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 z 0 z 1 Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 z 1 , we have
ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 ψ 0 P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 z 0 Γ P , Q 1 , , Q r 1 , S 1 , , S r 2 Γ P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 z 0 .
Finally we use in the last inequality the representation of the matrices in the form
P = 1 i j n p i j Δ i j , Q k = 1 i j n q i j , k Δ i j , S m = 1 i j n s i j , m Δ i j ,
and the linearity of Γ with respect to ( P , Q 1 , , Q r 1 , S 1 , , S r 2 ) , which allows us to get the presentation of the generalized gradient in the form (11). ☐
When solving (10), we pass from a constrained problem to an unconstrained one. We define the penalty functions
ψ 1 ( B ) = λ max ( B ) 1 , ψ 2 ( B ) = λ min ( B ) , B Ω n ,
and the corresponding Lagrange function
L ( P , Q 1 , , Q r 1 , S 1 , , S r 2 , u ) : = ψ 0 ( P , Q 1 , , Q r 1 , S 1 , , S r 2 ) + δ = 1 , 2 u P , δ ψ δ ( P ) + k = 1 r 1 u Q k , δ ψ δ ( Q k ) + m = 1 r 2 u S m , δ ψ δ ( S m ) ,
where u = { u P , 1 , u Q 1 , 1 , , u Q r 1 , 1 , u S 1 , 1 , , u S r 2 , 1 , u P , 2 , u Q 1 , 2 , , u Q r 1 , 2 , u S 1 , 2 , , u S r 2 , 2 } R 2 ( 1 + r 1 + r 2 ) are non-negative Lagrange multipliers.
Theorem 2.
Provided that the condition H1 holds, the function ψ 0 P , Q 1 , , Q r 1 , S 1 , , S r 2 attains its minimum at the point P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 i = 1 1 + r 1 + r 2 Ω ¯ n 1 if and only if the point
( P 0 , Q 1 , 0 , , Q r 1 , 0 , S 1 , 0 , , S r 2 , 0 , u 0 ) , where
u 0 = ( u P , 1 0 , u Q 1 , 1 0 , , u Q r 1 , 1 0 , u S 1 , 1 0 , , u S r 2 , 1 0 , u P , 2 0 , u Q 1 , 2 0 , , u Q r 1 , 2 0 , u S 1 , 2 0 , , u S r 2 , 2 0 ) ,
is a saddle point of the Lagrange function (12).
Proof. 
The objective function ψ 0 ( · ) and the constraint functions ψ 1 ( · ) , ψ 2 ( · ) are convex for the matrices P , Q 1 , , Q r 1 , S 1 , , S r 2 Ω n . It follows from the convexity of maximum eigenvalue and the concavity of minimum eigenvalue of a symmetric matrix (see Example 3.10 in [18]). When proving the convexity of the function ψ 0 with respect to P , Q 1 , , Q r 1 , S 1 , , S r 2 i = 1 1 + r 1 + r 2 Ω n we also use the linear dependence of the operator Γ on P , Q 1 , , Q r 1 , S 1 , , S r 2 .
Due to the Karush–Kuhn–Tucker conditions for convex problems it is left to show that Slater conditions hold [18] (page 244). By virtue the assumption H1 there exists a tuple of matrices P ¯ , Q ¯ 1 , , Q ¯ r 1 , S ¯ 1 , , S ¯ r 2 i = 1 1 + r 1 + r 2 Ω n dom ( ψ 0 ) such that
ψ δ ( P ¯ ) < 0 , ψ δ ( Q ¯ 1 ) < 0 , , ψ δ ( Q ¯ r 1 ) < 0 , ψ δ ( S ¯ 1 ) < 0 , , ψ δ ( S ¯ r 2 ) < 0 , δ = 1 , 2 ,
is satisfied, and the Slater condition is true, which concludes the proof. ☐

4. Conclusions

The work is devoted to modeling and investigation of the architecture design for the delayed recurrent neural network basing on the delayed differential equations. The usage of discrete and distributed delays makes it possible to model the calculation of the next states using internal memory, which corresponds to the artificial recurrent neural network architecture used in the field of deep learning.
The paper proposes a method for constructing an exponential estimate of the solution of a model of a recurrent neural network using the Lyapunov–Krasovskii functional. This estimate is reduced to solving the corresponding linear matrix inequality. This is the most costly operation in terms of the computational complexity, however, this is where we can apply the optimization approach to find the optimal set of matrices from the viewpoint of the exponential estimate.
In contrast to the indirect method of constructing exponential estimates, which was proposed in previous works [12,13], in this study, the method based on the Lyapunov–Krasovskii functional allows the optimization of the estimate within the compact domain of positive definite matrices.
To this end, the concept of a generalized gradient of a convex function on a set of positive definite matrices is introduced. The constructive form of the generalized gradient for the minimal eigenvalue of the matrix Γ is presented. The Lagrange function for the unconditional optimization problem is constructed. In this case, the search for the optimal exponential estimate is reduced to finding the saddle point of the Lagrange function.

Author Contributions

Conceptualization, V.M. and M.K.; methodology, S.R.; software, V.M.; validation, V.M., S.R. and M.K.; formal analysis, V.M.; investigation, S.R.; resources, M.K.; data curation, V.M.; writing—original draft preparation, S.R.; writing—review and editing, M.K.; visualization, V.M.; supervision, S.R.; project administration, M.K.; funding acquisition, M.K. All authors have read and agreed to the published version of the manuscript.

Funding

The APC was funded by the University of Bielsko-Biala.

Acknowledgments

The work is supported by the University of Bielsko-Biala, and the program number is K18/1b/UPBJ/2019-2020.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LSTMLong Short-Term Memory
GRUGated Recurrent Units
RNNRecurrent Neural Network
DDEDelayed Differential equarion

References

  1. Habiba, M.; Pearlmutter, B.A. Neural Ordinary Differential Equation based Recurrent Neural Network Model. arXiv 2020, arXiv:2005.09807. [Google Scholar]
  2. Sherstinsky, A. Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef] [Green Version]
  3. Fang, S.; Jiang, M.; Wang, X. Exponential convergence estimates for neural networks with discrete and distributed delays. Nonlinear Anal. Real World Appl. 2009, 702–714. [Google Scholar] [CrossRef]
  4. Hale, J.K.; Lunel, S.M.V. Introduction to Functional Differential Equations; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013; Volume 99. [Google Scholar]
  5. Zhang, H.; Wang, Z.; Liu, D. A Comprehensive Review of Stability Analysis of Continuous-Time Recurrent Neural Networks. IEEE Trans. Neural Networks Learn. Syst. 2014, 25, 1229–1262. [Google Scholar] [CrossRef]
  6. Hopfield, J.J. Neurons with graded response have collective computational properties like those of two-state neurons. Proc. Natl. Acad. Sci. USA 1984, 81, 3088–3092. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Wei, J.; Ruan, S. Stability and bifurcation in a neural network model with two delays. Phys. D Nonlinear Phenom. 1999, 130, 255–272. [Google Scholar] [CrossRef]
  8. Ruan, S.; Wei, J. On the zeros of a third degree exponential polynomial with applications to a delayed model for the control of testosterone secretion. Math. Med. Biol. 2001, 18, 41–52. [Google Scholar] [CrossRef]
  9. Ruan, S.; Wei, J. On the zeros of transcendental functions with applications to stability of delay differential equations with two delays. Dyn. Contin. Discret. Impuls. Syst. Ser. A 2003, 10, 863–874. [Google Scholar]
  10. Yan, X.P.; Li, W.T. Stability and bifurcation in a simplified four-neuron BAM neural network with multiple delays. Discret. Dyn. Nat. Soc. 2006, 2006. [Google Scholar] [CrossRef]
  11. Huang, C.; Huang, L.; Feng, J.; Nai, M.; He, Y. Hopf bifurcation analysis for a two-neuron network with four delays. Chaos Solitons Fractals 2007, 795–812. [Google Scholar] [CrossRef]
  12. Martsenyuk, V. On an indirect method of exponential estimation for a neural network model with discretely distributed delays. Electron. J. Qual. Theory Differ. Equ. 2017, 2017, 1–16. [Google Scholar] [CrossRef]
  13. Martsenyuk, V. Indirect method of exponential convergence estimation for neural network with discrete and distributed delays. Electron. J. Differ. Equ. 2017, 2017, 1–16. [Google Scholar]
  14. Gu, K. An integral inequality in the stability problem of time-delay systems. In Proceedings of the 39th IEEE Conference on Decision and Control (Cat. No.00CH37187), Sydney, Australia, 12–15 December 2000. [Google Scholar] [CrossRef]
  15. Sanchez, E.; Perez, J. Input-to-state stability (ISS) analysis for dynamic neural networks. IEEE Trans. Circuits Syst. Fundam. Theory Appl. 1999, 46, 1395–1398. [Google Scholar] [CrossRef]
  16. Zhang, F. (Ed.) The Schur Complement and Its Applications; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar] [CrossRef]
  17. Khusainov, D.Y.; Martsenyuk, V.P. Optimization method for stability analysis of delayed linear systems. Cybern. Syst. Anal. 1996, 32, 534–538. [Google Scholar] [CrossRef]
  18. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Martsenyuk, V.; Rajba, S.; Karpinski, M. On Optimization Techniques for the Construction of an Exponential Estimate for Delayed Recurrent Neural Networks. Symmetry 2020, 12, 1731. https://doi.org/10.3390/sym12101731

AMA Style

Martsenyuk V, Rajba S, Karpinski M. On Optimization Techniques for the Construction of an Exponential Estimate for Delayed Recurrent Neural Networks. Symmetry. 2020; 12(10):1731. https://doi.org/10.3390/sym12101731

Chicago/Turabian Style

Martsenyuk, Vasyl, Stanislaw Rajba, and Mikolaj Karpinski. 2020. "On Optimization Techniques for the Construction of an Exponential Estimate for Delayed Recurrent Neural Networks" Symmetry 12, no. 10: 1731. https://doi.org/10.3390/sym12101731

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop