Existence of Periodic Solutions for an Output Hidden Feedback Elman Neural Network ()
1. Introduction
Artificial neural networks are computational paradigms which implement simplified models of their biological counterparts, biological neural networks. Biological neural networks are the local assemblages of neurons and their dendritic connections that form the (human) brain. Accordingly, artificial neural networks are characterized by
· local processing in artificial neurons (or processing elements);
· massively parallel processing, implemented by rich connection pattern between processing elements;
· the ability to acquire knowledge via learning from experience;
· knowledge storage in distributed memory, the synaptic processing element connections.
Neural networks process information in a similar way the human brain does. The network is composed of a large number of highly interconnected processing elements (neurons) working in parallel to solve a specific problem. Neural networks learn by example.
An important application of neural networks is pattern recognition. Pattern recognition can be implemented by using a feed-forward neural network that has been trained accordingly. During training, the network is trained to associate outputs with input patterns. When the network is used, it identifies the input pattern and tries to output the associated output pattern. The power of neural networks comes to life when a pattern that has no output associated with it, is given as an input. In this case, the network gives the output that corresponds to a taught input pattern that is least different from the given pattern.
Elman neural network [1] is a kind of recurrent neural network. Compared with traditional neural networks, an Elman neural network has additional inputs from the hidden layer, which forms a new layer—the context layer. So the standard back-propagation algorithm used in Elman neural network is called Elman back-propagation algorithm. Elman neural network can be applied to solve prediction problems of discrete time sequences [2] [3] [4].
The Elman neural network is one of the most widely used and most effective neural network models in artificial neural networks and has powerful processing ability for nonlinear decisions [5] [6]. The Elman neural network can be considered as a special kind of feed forward neural network with additional memory neurons and local feedback. Because of its better learning efficiency, approximation ability, and memory ability than other neural networks, the Elman neural network can not only be used in time series prediction, but also in system identification and prediction [4] [7] [8] [9] [10].
Shi et al. [11] proposed Output Hidden Feedback (OHF) Elman neural network based on modified Elman neural network [7] by introducing a feedback between the output layer and an additional, output context layer.
The existence of periodic solutions is a classical problem of the qualitative theory of differential and difference equations. Numerous papers have been devoted to the existence of periodic solutions of different kinds of neural networks with continuous and discrete time. In [12], sufficient conditions were found for the existence and global exponential stability of a class of Hopfield neural networks with periodic impulses and finite distributed delays. In [13], the authors found sufficient conditions for the global exponential periodicity of a discrete-time counterpart of a bidirectional associative memory neural network. In [14], sufficient conditions are obtained for the existence and global asymptotic stability of periodic solutions for delayed complex-valued simplified Cohen-Grossberg neural networks.
In [15] [16], for two different classes of Hopfield-type neural networks with periodic impulses and finite distributed delays we introduced discrete-time counterparts. Using different methods, we found sufficient conditions for the existence and global exponential stability of a unique periodic solution of the discrete systems considered. In [17], sufficient conditions were found for the existence of periodic solutions for the discrete-time counterpart of a neutral-type cellular neural network with time-varying delays and impulses. In [18], we found sufficient conditions for the existence of periodic solutions for the discrete-time counterpart of a complex-valued Hopfield neural network with time-varying delays and impulses. In [19], we proved the global exponential periodicity of a class of Hopfield neural networks with distributed delays and impulses. In our recent paper [20], we obtained sufficient conditions for the existence of a periodic output of a modified Elman neural network with a periodic input by using Mawhin’s continuation theorem of coincidence degree theory [21].
In the present paper, we consider an OHF Elman neural network [4] with a periodic input. Using the result of [20], we find sufficient conditions for the existence of a periodic output of the neural network considered. Furthermore, for a subclass of these OHF Elman neural networks, we shall find the periodic output in a straightforward way using another sufficient condition. Examples illustrating these sufficient conditions are given. The calculations are done using MATLAB.
2. Preliminaries: Modified Elman Neural Network
Here we recall the results of our paper [20]. We consider a modified Elman neural network with r nodes in the input layer, n nodes in the hidden and context layers, respectively, and m nodes in the output layer, which adds a self-feedback factor α,
, in the context nodes, based on the traditional Elman neural network [4] [10]. Its mathematical model is:
(1)
(2)
(3)
Here
is the set of all positive integers, the input u is an r-dimensional vector, the output x of the hidden layer and the output
of the context nodes are n-dimensional vectors, while the output y of the output layer is an m-dimensional vector. The weights
,
, and
respectively of the context nodes, input nodes and hidden nodes are entries of
-,
- and
-dimensional matrices, respectively;
,
, are the transfer functions of the hidden layer often taken as sigmoid functions,
,
, are the transfer functions of the output layer and are often taken as linear functions. An Elman neural network with
,
and
is depicted in Figure 1. The values of the numbers r, n and m in Figure 1 and Figure 2 and Examples 1 and 2 are chosen quite small for the sake of simplicity.
Clearly, for a given input
,
, and initial values
,
, we can find the output
,
, from Equations (1)-(3).
Now suppose that the input
is N-periodic for some
, that is,
,
. We shall look for sufficient conditions for the existence of an N-periodic output
,
. This means that, for a suitable
Figure 1. An Elman neural network with
,
and
.
choice of the initial values
,
, the output
is N-periodic. For this purpose, it suffices that the output
of the hidden layer is N-periodic. From now to the end of the present section, we restrict our attention to Equations (1), (2).
Further on, for convenience, we consider Equations (1), (2) for
, that is,
(4)
(5)
where
is the set of all integers. We assume that
for
. Sufficient conditions for the existence of an N-periodic solution
,
, of Equations (4), (5) are given below.
We make the following assumptions:
A1.
.
A2. There exist positive constants
,
, such that
for all
,
. (6)
A3. There exists a positive integer N such that
for all
,
. (7)
A4.
.
In order to formulate our main result, we introduce the
-matrix
(8)
where
is the Kronecker delta, and assume that:
A5. The matrix
is an M-matrix.
Assumption A5 implies that the matrix
is nonsingular and its inverse has only nonnegative entries [22] [23].
The main result of [20] is the following theorem.
Theorem 1. Suppose that assumptions A1-A5 hold. Then the system of Equations (4), (5) has at least one N-periodic solution
.
Theorem 1 is proved using Mawhin’s continuation theorem [ [21], p. 40].
Example 1. Consider a modified Elman neural network with
,
and
(as in Figure 1). Suppose that the transfer functions
,
, of the hidden layer all equal the sigmoid function
,
,
,
are arbitrary N-periodic functions for some positive integer N, say,
(9)
where
is the greatest integer in the real number
, that is,
(10)
(11)
(a) Further on, let us assume that the weights
,
, of the context nodes all equal
, the transfer functions
,
, the weights
,
, of the input nodes and
,
, of the hidden nodes are arbitrary.
Then, assumption A2 is satisfied with
,
, assumption A4 is also satisfied since
(12)
Finally, the matrix
(13)
is an M-matrix with inverse
(14)
Since all assumptions of Theorem 1 are satisfied, the modified Elman neural network under consideration has an N-periodic output
,
.
Let us assume that, moreover,
,
, and
. Then the system of Equations (4), (5) takes the form
(15)
(16)
where
(17)
For
and
, we recall that
(k and
are congruent modulo N) if and only if
. For instance,
means that
for some
.
It suffices to find the initial values
,
,
, so that
(18)
Equations (15), (16) imply that
,
, and
,
Thus, in order to satisfy Equations (18), the initial conditions must be chosen so that
and
. The system of Equations (15), (16) reduces to
(19)
(20)
We have found that the initial values
,
satisfying Equations (18) are (approximately)
,
. The first 4 values of the 3-periodic solution of Equations (15), (16) are presented in Table 1.
This solution can be found with arbitrarily high accuracy.
(b) Next, let us assume that
,
,
,
Table 1. A 3-periodic solution of Equations (15), (16).
,
, the transfer functions
,
, the weights
,
, of the input nodes and
,
, of the hidden nodes are still arbitrary.
Then, assumption A2 is still satisfied with
, assumption A4 is also satisfied since
(21)
Finally, the matrix
(22)
is an M-matrix with inverse
(23)
Since all assumptions of Theorem 1 are satisfied, the modified Elman neural network under consideration has an N-periodic output
,
.
Let us assume that, moreover,
,
,
,
, and
. Then, the system of Equations (4), (5) takes the form
(24)
(25)
(26)
(27)
It suffices to find the initial values
, so that Equations (18) are satisfied. We have found that, approximately,
,
,
,
,
,
(see Table 2 for the first 4 values of the 3-periodic solution of Equations (24)-(27)).
The initial values in Example 1, (a) and (b) have been found after numerous experiments with different sets of possible initial values, using MATLAB. They can be found with an arbitrarily high accuracy after sufficiently many iterations.
3. Output Hidden Feedback Elman Neural Networks: Main Results
OHF Elman neural network achieves the ability to process dynamic data by adding feedback from the output layer to the hidden output context layer (second context layer) based on Elman neural network. The mathematical model of an OHF Elman neural network is [4]:
(28)
(29)
(30)
(31)
where Equations (28), (29) are the same as Equations (1), (2), the transfer functions
,
, of the output layer are as in Equation (3),
is the gain factor of the self-feedback of the output layer,
are the connection
Table 2. A 3-periodic solution of system (1), (2) in Case (b).
Figure 2. An OHF Elman neural network with
,
and
.
weights of the second context layer nodes, and
is the output of the second context layer. An OHF Elman neural network with
,
and
is depicted in Figure 2.
Clearly, for a given input
,
, and initial values
,
,
,
, we can find the output
,
, from Equations (28)-(31).
Now suppose that the input
is N-periodic for some
, that is,
,
. We shall look for sufficient conditions for the existence of an N-periodic output
,
. This means that for a suitable choice of the initial values
,
,
,
, the output
is N-periodic.
In addition to assumptions A1-A5, we make the following assumptions:
A6.
.
A7. There exist positive constants
,
, such that
for all
,
. (32)
A8.
.
In order to formulate our main result, we introduce the
-matrix
(33)
and assume that:
A9. The matrix
is an M-matrix.
Now, we can state our main result as the following theorem.
Theorem 2. Suppose that assumptions A1-A9 hold.Then the system of Equations (28)-(31) has at least one N-periodic solution
.
Proof. According to Theorem 1, the system of Equations (28), (29) has an N-periodic solution
. Then the system of Equations (30), (31) with N-periodic input
is of the form of Equations (1), (2), thus it has at least one N-periodic solution
.¨
As mentioned in Section 2, the transfer functions
,
, of the output layer are often taken as linear functions. Without loss of generality, we can assume that:
A10.
.
where
are some constants. In this case, in the assumptions of Theorem 2 we have
.
Now we show that, once the N-periodic solution
of the system of Equations (28), (29) has been found, in the case of linear transfer functions of the output layer the N-periodic solution
can be found in a straightforward way using another sufficient condition.
For convenience, we introduce the matrices
(34)
is the
unit matrix. Then Equations (30), (31) can be written in a matrix form as
(35)
We successively obtain
(36)
(37)
(38)
(39)
(40)
By induction, we prove that
(41)
(42)
where
are some linear functions of their arguments.
In order to obtain an N-periodic solution
of the system of Equations (30), (31) we need to find initial conditions
satisfying
(43)
From Equations (41)-(43) we derive
(44)
where
(45)
If
A11. The
matrix
is nonsingular, then from Equation (44) we can determine the initial values
. Thus, we have proved.
Theorem 3. Suppose that assumptions A1-A6, A10, A11 hold. Then the system of Equations (28)-(31) has at least one N-periodic solution
.
Example 2. Now let us consider an OHF Elman neural network given by Equations (28)-(31), with
,
and
,
, where Equation (28), (29) are as in Example 1(b),
, assumption A7 is satisfied with
,
, the matrix of the connection weights of the nodes of the second context layer is
(46)
the matrix C of the weights of the hidden layer is arbitrary.
Assumption A8 is satisfied since
(47)
The matrix
(48)
is an M-matrix with inverse
(49)
Since all assumptions of Theorem 2 are satisfied, the OHF Elman neural network under consideration has a 3-periodic output
.
Now let also assumption A10 be satisfied with still arbitrary
,
. Equation (44) takes the form
(50)
where
(51)
For the
blocks of this matrix we find
(52)
(53)
(54)
(55)
We give the entries of the
matrix
in the form of Table 3.
Clearly, the matrix
is nonsingular. By virtue of Theorem 3, the system of Equations (28)-(31) has a 3-periodic solution
. Moreover, for a 3-periodic solution
,
, of Equations (28), (29), and given a
weight matrixC of the hidden layer and a constant vector
, we can express the initial conditions
from Equation (50) making use of the inverse matrix
.
Now, let us recall the 3-periodic solution
,
, of the system of Equations (28), (29):
Table 3. Entries of the matrix
.
(56)
We assume that
(57)
and evaluate the right-hand side of Equation (50):
(58)
(59)
Now from Equation (50) we find
(60)
Table 4. A 3-periodic solution of the system of Equations (28)-(31).
Finally, we find the first four values of the approximate 3-periodic solution (see Table 4). We notice that Equations (43) are satisfied much more precisely than Equations (18) in Example 1. On the other hand, the application of the method becomes much more difficult for a greater number m of the nodes in the output layer.
4. Conclusion
We presented sufficient conditions for the existence of periodic output of modified and OHF Elman neural networks with periodic input. Examples illustrating the results obtained were given. The models considered can be applied to the services quality of experience prediction.