Distributed Learning for Echo State Networks with Dynamic Event-Triggered Consensus

Abstract

This paper is devoted to studying the dynamic event-triggered consensus problem in distributed Echo State Networks (ESNs). The traditional ESN problem can be formulated as a set of distributed sub-problems with consensus constraints, which are solved using the Zero-Gradient-Sum (ZGS) distributed optimization strategy. Additionally, each agent employs a distributed dynamic event-triggered strategy based on its internal dynamic variables to achieve asymptotic consensus convergence. The purposeful dynamic triggering strategy is more communication-efficient compared to static triggering strategies and eliminates the need for continuous communication to update controllers or monitor triggering thresholds. Numerical simulation results are presented to validate the efficacy of the proposed algorithm.

Share and Cite:

Gong, X. and Ai, W. (2025) Distributed Learning for Echo State Networks with Dynamic Event-Triggered Consensus. Journal of Computer and Communications, 13, 35-49. doi: 10.4236/jcc.2025.134003.

1. Introduction

Echo state network is a novel type of recurrent neural network. Compared to other recurrent neural networks, ESN can handle time series classification and regression tasks by simply training the weight matrix connecting the hidden layer to the output layer using linear regression [1]. In practical applications, due to the large volume of data and privacy concerns, distributed ESNs have been widely applied [2]. In recent years, extensive research has been conducted on ESNs. For example, Zhu proposed a distributed learning method for DeepESN (DESN) based on improved singular value decomposition (SVD) and alternating direction method of multipliers (ADMM), which effectively reduces computation and communication costs for training DESNs in edge environments [3]. In [4], a novel distributed ESN model integrated with an auto-encoder (AE-DESNm) eliminates the dimension disaster problem and uses an extreme learning machine (ELM) to extract features for estimating key variables. Integrating ESN with the decentralized average consensus (DAC) [5] approach facilitates achieving consensus in output values. Over time, considerable advancements have been made in methods for enhancing ESNs. For instance, a hybrid evolutionary algorithm that combines a competitive swarm optimizer with local search has been applied to mitigate the challenges associated with the random selection of input and reservoir weights [6]. Additionally, an optimized echo state network (O-ESN) proposed in [7] utilizes binary particle swarm optimization to minimize network complexity while improving generalization capabilities.

In a distributed framework, the assumption for achieving consensus convergence is often that each agent communicates continuously with its neighbors, which typically requires unlimited communication resources. However, a perfect communication network with infinite bandwidth capacity is usually not feasible in real-life scenarios [8]. To address the resource consumption issue in continuous communication among multi-agents, Lu proposed a general sampling framework where the sampled signals originate from a known union of subspaces, and the sampling operator is linear. Furthermore, the framework identified the minimum sampling requirements for several classes of signals [9]. If the continuous sampling signals fluctuate minimally, agents will engage in continuous unnecessary communication, resulting in a large amount of wasted samples [10]. The introduction of the event-based control method has garnered widespread interest, as it adjusts the sampling sequence based on events related to the system state, thus avoiding the communication resource wastage associated with traditional periodic sampling [11]. An event-triggered consensus protocol and triggering law proposed in [12] can achieve global consensus if and only if the underlying directed graph has a directed spanning tree. Moreover, it is found that the existence of a directed spanning tree is always a necessary and sufficient condition for achieving global consensus, regardless of whether input saturation is present. This conclusion remains valid even when more complex nonlinear dynamic behaviors are introduced. The fully distributed event-triggered algorithm proposed by Berneberg ensures asymptotic convergence to the average consensus state while allowing each agent to independently select a desired minimum inter-event time, thereby guaranteeing non-Zeno behavior [13]. In [14], static and dynamic event-triggered methods are introduced as secondary control techniques to restore voltage frequency and improve power sharing accuracy.

A new dynamic event-triggered control strategy, which includes internal dynamic variables, was proposed in [15]. Compared to static event-triggered strategies, it offers more advantages. As a result, dynamic event-triggered strategies have been used in recent years to solve the consensus problem in multi-agent systems (MAS). Later, dynamic event triggering was applied to islanded DC microgrids [16] and vehicle platoon systems [17], where the data from neighboring DGs are used only at event-triggered times, thus reducing the burden on the communication network. In our previous work [18], we applied an event-triggered control policy to the communication component of the problem to avoid unnecessary transmissions, where the agents send messages only when the negotiation is critically needed. In [19], a distributed cooperative learning algorithm designed for stochastic configuration networks with fixed-time convergence, known as FixD-SCN, which is presented to tackle the ‘Big Data’ problem. To our best knowledge, little effort has been made in the field of dynamic event-triggered distributed learning over peer-to-peer networks, especially for RNNs.

The purpose of this paper is to propose a distributed learning algorithm for training the ESN, in which all agents converge asymptotically while minimizing the use of communication resources during the process. To accomplish this objective, we extend the multi-agent system with dynamic event-triggered algorithms to address the distributed learning problem. Initially, we convert the centralized ESN into a distributed form by breaking down the training process into smaller components, each subject to equality constraints on the output parameters. Next, a dynamic event-triggered algorithm is used to reduce communication during the process of optimizing the distributed ESN output weights. The proposed distributed algorithm is inspired by the use of dynamic event-triggered techniques to achieve distributed consensus convergence [20]. This work extends that of Scardapane et al. [21] by computing the final optimal consensus of ESN output parameters. The primary contributions presented in this paper are summarized below.

  • A fully decentralized algorithm for training ESNs is introduced, where the training process is collaboratively executed without relying on a central coordinator.

  • The convergence with dynamic event-triggered mechanism of a distributed learning algorithm is addressed to provide an asymptotically convergent.

  • The convergent property of the proposed algorithm is mathematically substantiated when the Lyapunov theory is employed.

Notation: Denote the set of real (natural) numbers as ( ). Let 1 m ( 0 m ) represent the m -dimensional column vector where each entry is 1 (0). The matrix I L L×L is the identity matrix. Letting and 1 represent the Euclidean norm (2-norm) and 1-norm, respectively. ρ( ) denotes the spectral radius of a matrix. is used to denote the transpose of a vector or matrix and to represent the Kronecker product. For a vector x= [ x 1 , x 2 ,, x D ] D , | x |= [ | x 1 | ,, | x D | ] , denote the υ -norm of x by x υ = ( i=1 D | x i | υ ) 1 υ , denote the sign of x by sign( x )= [ sign( x 1 ),,sign( x D ) ] , where sign( ) represents the sign function; denote x [ α ] = [ x 1 [ α ] ,, x D [ α ] ] with x i [ α ] = | x i | α sign( x i ) , where | | denotes the absolute value. Then we have x i sign( x i )=| x i | , x i x i [ α ] = | x i | α+1 and | x i | x i [ α ] = x i [ α+1 ] . Let _ denote the matrix inequality sign, i.e., A _ B , indicating that BA is a positive semidefinite matrix.

2. Problem Formulation

2.1. Graph Theory

We design a communication network of J agents distributed over a geographic area. Mathematically, we represent the communication network by a graph G=( V,,A ) of order J with V={ 1,,J } depicting the set of vertices, V×V representing the edge set, and A=[ a ij ] J×J illustrates the weighted adjacency matrix, where a ij >0 if ( j,i ) and a ij =0 , otherwise. The graph has no loops or multiple edges. If there is a connection between node i and node j , ( j,i ) . The set of neighbors of node i is represented by N i ={ j:( j,i ) } . Define the Laplacian matrix =[ l ij ] J×J , where l ij = a ij for ij and l ii = j=1 J a ij . The matrix is symmetric and positive semi-definite, with all eigenvalues being non-negative.

2.2. Problem Formulation

Lemma 1 (graphs) For an undirected connected graph G with J vertices, the eigenvalues of contains only one zero entry. 0= λ 1 ( ) possessing a right eigenvector, 1 , i.e., 1=0 , and other have positive real parts. Suppose 0= λ 1 ( )< λ 2 ( ) λ J ( )= λ max ( ) in an ascending order. The minimum nonzero eigenvalue λ 2 ( )= min x0, 1 x=0 x x x x , where x= [ x 1 ,, x J ] . We have

x x= 1 2 i=1 J j=1 J a ij ( x j x i ) 2 . (1)

Furthermore, if 1 x=0 , then x x λ 2 ( ) x x .

Lemma 2 (nonsingular) Let κ 1 , κ 2 ,, κ J 0 . Then

( i=1 J κ i ) p i=1 J κ i p J 1p ( i=1 J κ i ) p (2)

if 0<p1 , and

J 1p ( i=1 J κ i ) p i=1 J κ i p ( i=1 J κ i ) p (3)

if 1<p< .

2.3. Distributed ESN over Distributed Datasets in a Multi-Agent Networks

ESN consists of three parts: an input layer, a hidden layer (reservoir), and an output layer. It is a variant of feedforward neural networks, with the training process involving feeding back output results to the hidden layer to adjust connection weights. The input vector x n D in , with a dimensionality of D in , is fed into a reservoir of dimensionality D res . The internal state of the reservoir, g n D res , is computed as follows:

g n = f res ( β in res x n + β res res g n1 + β out res y n1 ), (4)

where β in res D res × D in , β res res D res × D res and β out res D res × D out are matrices generated randomly. The nonlinear function f res ( ) is applied, where y n1 D out denotes the output from the previous time step. The output at the current step is then calculated as

y n = f out ( β in out x n + β res out g n ). (5)

Depending on the specific dataset, the parameters β in out D out × D in and β res out D out × D res are adapted, so as an invertible nonlinear function f out ( ) . To train the readout, assume that there is a sequence of N input-output pairs ( x 1 , y 1 ),,( x N , y N ) . The inputs are mapped to the reservoir, generating a series of internal states g 1 ,, g N . At this stage, due to the unavailability of the ESN output for feedback, the designated output is substituted in (4). We obtain the hidden matrix H and output matrix S as

H=[ x 1 , g 1 x N , g N ],S=[ f out 1 ( y 1 ) f out 1 ( y N ) ]. (6)

The optimal weight vector for the output is obtained by solving the subsequent regularized least-squares problem:

β * = argmin β { 1 2 HβS 2 + C 2 β 2 }, (7)

where β= [ β in out , β res out ] ( D in + D res )× D out and C>0 is a regularization parameter. The solution to the problem in (7) can be written as

β * = ( H H+C I L ) 1 H S, (8)

where L= D in + D res .

To transform the centralized ESN into a distributed ESN, at the outset, we break down the centralized ESN task into several subtasks, each corresponding to a specific agent. Obtaining the optimal solution for the regularized problem leads to the determination of the output weights β * .

min { β i },{ ξ n,i } 1 2 i=1 J n=1 N i ξ n,i 2 + C 2J i=1 J β i 2 s.t. β i h n,i = s n,i ξ n,i ,iV,n=1,, N i , β i = β j ,iV,j N i . (9)

The training error vector for the output neurons, corresponding to the training example x n,i at the i -th learning agent, is denoted by ξ n,i D out . To convert a centralized ESN into a fully distributed ESN, the overall output weights β are first duplicated and assigned to each agent. Equality constraints are then enforced along the edges of the communication network. By incorporating the constraints into the objective function, problem (9) can be reformulated as

min { β i } i=1 J ( 1 2 H i β i S i 2 + C 2J β i 2 ) s.t. β i = β j ,iV,j N i , (10)

where the hidden matrix H i and the training target matrix S i are autonomously acquired from the respective dataset D i observed by agent i itself. For simplicity in presentation, we begin by simplifying problem (10) as

β * = argmin β { i=1 J u i ( β )u( β ) }, (11)

where

u i ( β )= 1 2 H i β S i 2 + C 2J β 2 (12)

is the individual function performed by agent i locally. Agent i is capable of independently optimizing its local objective and establishing local consensus constraints through communication with neighboring agent j .

2.4. Distributed ESN Solved by the ZGS Algorithm

The ZGS algorithm aims to achieve optimal consensus on a manifold where the sum of the gradients of local functions is consistently zero. Let β i ( t ) denote the state of agent i estimating the unknown optimum β * at iterative step k , with β i ( 0 ) indicating the initial value. The objective is to drive β i ( t ) of each agent asymptotically to β * , i.e., assuming convexity in the objective function, the global objective can be optimally achieved uniquely. According to the form of the objective function, it is known that it is a convex function. The global objective can be uniquely achieved optimally.

Assumption 1 For each iV , the function u i demonstrates strong convexity and maintains twice continuous differentiability.

Under Assumption 1, there is a unique minimizer β i * , i.e.,

β i * = ( H i H i + C J I L ) 1 H i S i (13)

such that the gradient u i ( β i * )=0 . The ZGS algorithm is formulated as an iterative procedure under the assumption of a zero gradient for u i . The following implicit distributed algorithm is considered to achieve the goal.

u i ( β i ( t+1 ) ) u i ( β i ( t ) ) = γ 1 j N i a ij ( β j ( t ) β i ( t ) ) [ α 1 ] + γ 2 j N i a ij ( β j ( t ) β i ( t ) ) [ α 2 ] (14)

with β i ( 0 )= β i * , where γ 1 , γ 2 >0,0< α 1 <1, α 2 >1 and a ij represents entries in the weighted adjacency matrix A of graph G . Under the assumption that the graph G is undirected and connected, we have

i=1 J ( u i ( β i ( t+1 ) ) u i ( β i ( t ) ) )=0. (15)

Assuming β i ( 0 )= β i * , we have

i=1 J u i ( β i ( 0 ) )= i=1 J u i ( β i * )=0. (16)

Consequently, the consensus value β o must be the optimizer β * . In such sense, a shared β * is determined for each agent, ensuring u( β * )= i=1 J u i ( β * )=0 . So that (10) holds. Furthermore, the gradient of the function u i ( β ) with respect to β can be easily computed as u i ( β )=( H i H i + C J I L )β H i S i . The Hessian of u i ( β ) is

2 u i ( β )= H i H i + C J I L Ω ˜ i . (17)

The matrix Ω ˜ i is static, nonsingular, symmetric and positive definite. Letting θ i = λ min ( Ω ˜ i )>0 and Θ i = λ max ( Ω ˜ i )>0 denote the minimum and maximum eigenvalues of Ω ˜ i , respectively, then

θ i I L _ Ω ˜ i _ Θ i I L . (18)

From the gradient of u i ( β ) , we have u i ( β i ( t+1 ) ) u i ( β i ( t ) ) = Ω ˜ i ( β i ( t+1 ) β i ( t ) ) . Then, β i ( t+1 ) β i ( t )= Ω ˜ i 1 ( u i ( β i ( t+1 ) ) u i ( β i ( t ) ) ) . Substituting this expression into the left-hand side of (14), an algorithm that iteratively solves the problem in a distributed manner is designed. For each agent iV , we design an algorithm for distributed ESN (ZGS-ESN). The ZGS-ESN iterates as follows:

β i ( t+1 )= β i ( t )+ γ 1 Ω ˜ i 1 j N i a ij ( β j ( t ) β i ( t ) ) [ α 1 ] + γ 2 Ω ˜ i 1 j N i a ij ( β j ( t ) β i ( t ) ) [ α 2 ] (19)

where γ 1 , γ 2 >0,0< α 1 <1, α 2 >1 , with initial values

β i ( 0 )= Ω ˜ i 1 H i S i , (20)

where the matrix Ω ˜ i is defined in (17). However, the continuous updating strategy may result in elevated communication consumption and frequent updates. The dynamic event-triggered strategy is an effective approach to overcome these disadvantages, in which the information interactions only happen at the trigger instants.

3. Distributed ESN with Dynamic Event-Triggered Consensus

Under the dynamic event-triggered strategy, we consider the following distributed algorithm:

β i ( t+1 )= β i ( t )+ Ω ˜ i 1 j N i a ij ( γ 1 ( β j ( t k i ) β i ( t k i ) ) [ α 1 ] + γ 2 ( β j ( t k i ) β i ( t k i ) ) [ α 2 ] ),t[ t k i , t k+1 i ) (21)

with initial values as in (20), where t k i is the latest triggering time for agent i . Drawing inspiration from [24], we begin by introducing the following combined measurement variable:

p i ( t )= j=1 N a ij ( β j ( t ) β i ( t ) ). (22)

Based on the definition of the combined measurement, we express the measurement error as follows :

e i ( t )= p i ( t k i ) p i ( t ). (23)

Agent i requires the combined measurement p i ( t ) only at specific time intervals, known as triggering times (or event times), which are denoted by t 0 i , t 1 i , t 2 i , . At each triggering time t k i , agent i must calculate p i ( t k i ) through communication with its neighbors. Given the current triggering time t k i , the subsequent triggering time t k+1 i is determined by the following triggering mechanism:

{ t k+1 i =inf{ t> t k i | h i d ( e i ( t ), j N i x j ( t ), η i ( t ) )0 }, η ˙ i ( t )= g i ( e i ( t ), p i ( t ), η i ( t ) ), (24)

where η i ( t ) represents the internal dynamic state and is the measurement error e i ( t ) that needs to be defined.

The triggering function h i d ( ) is affected not only by the state of the controlled system but also by the internal dynamic state, denoted as η i ( t ) . This type of triggering mechanism, referred to as a dynamic event-triggering mechanism, was first introduced for single controlled system in [15]. When the internal dynamic state η i ( t ) is disregarded , the dynamic event-triggering approach simplifies to a static event-triggering mechanism, expressed as h i s ( e i ( t ), j N i β j ( t ) )0 . In this study, the dynamic triggering conditions are defined by the following equations.

{ t k+1 i =inf{ t> t k i | e i ( t ) 2 δ i p i ( t ) 2 π i η i ( t )0 }, η ˙ i ( t )= τ i η i ( t )+ θ i ( δ i p i ( t ) 2 e i ( t ) 2 ), η i ( 0 )>0, (25)

where, θ i =68 , τ i =0.03 , π i =0.002 and δ i =0.01 . In a multi-agent system, both static and dynamic event-triggering mechanisms need to be decentralized, meaning that each triggering mechanism is restricted to accessing information from its neighboring agents.

4. Main Results

In this section, a numerical example is provided to verify the effectiveness of the proposed algorithm.

4.1. Description of the Datasets

The ZGS-FxTdt-ESN algorithm was validated on synthetic dataset designed for nonlinear system identification and the prediction of chaotic time series. For a large-scale simulation analysis, we utilize datasets that are approximately one to two orders of magnitude larger than those used in earlier studies. To be specific, for each dataset, we create 50 sequences, each comprising 2,000 elements. These sequences start from varying initial conditions, resulting in a total of 100,000 samples for each experiment.

The Mackey-Glass chaotic time-series (referred to as MKG) dataset is used for prediction tasks. Characterized in continuous time, this series follows the subsequent differential equation:

x ˙ n =0.1 x n + 0.2 x nτ 1+ x n30 10 . (26)

For τ>16.8 , the time-series given by (26) undergoes integration using a fourth-order Runge-Kutta method with a time step of 0.1. Subsequently, it is sampled every 10 time-instants. The task then becomes a 10-step-ahead prediction problem

y n = x n+10 . (27)

The functional form of Geometric Brownian Motion (GBM) is: S t = S t1 e ( μ 1 2 σ 2 )dt+σdW . In this model: μ=0.0003 represents the drift rate, governing the long-term trend of the asset price; σ=0.005 denotes the volatility, determining the magnitude of price fluctuations; S( t ) is the asset price at time t ; dW( t ) is the increment of Brownian motion over an infinitesimal time interval dt . Based on this model, we will generate 50 time series, each containing 2000 consecutive time steps of simulated asset price data.

4.2. The Setting of the Algorithms

We create an agent network employing a model with a randomly generated topology with a 55% connectivity rate. Our experimentation involves varying the number of agents, starting from 4 and incrementing by 4 up to 20. To assess testing errors, we employ a 3-fold cross-validation on the initial set of 50 sequences. Each sequence is 2000 units in length. In every fold of the cross-validation process, the training sequences are distributed across the agents, after which we assess the performance of the four algorithms listed below:

  • C-ESN: This algorithm is a single-agent ESN. The training data is centralized, and the ESN is trained by directly addressing the problem.

  • L-ESN: In this algorithm, each agent independently trains a localized ESN using its own data, with no communication taking place. The average testing error is computed across all agents.

  • ZGS-ESN: This algorithm set γ 2 =0 and α 1 =1 . The maximum number of iterations is 400 with an error level ϵ= 10 6 . The parameter γ 1 =0.01 . ZGS-FxTet-ESN: This algorithm is an extension of the ZGS-FxT-ESN with ω=0.01 . We set the maximum number of iterations at 400, and the error threshold at ϵ= 10 6 .

  • ZGS-FxTdt-ESN: We set the maximum number of iterations at 400, ω=0.01 and the error threshold at ϵ= 10 6 .

Each of the algorithms employs an identical ESN architecture, which is detailed in the following subsection. The threefold cross-validation process is iterated 10 times with variations in ESN initialization and data partitioning . Errors from each iteration and fold are accumulated. The trained ESN is applied to the test sequences to compute the error , and the resulting predicted outputs y ˜ 1 ,, y ˜ N test are compiled. In this context, N test denotes the count of testing samples excluding the initial washout elements from the test sequences. The normalized root mean-square error (N-RMSE) is defined as follows:

(a) [MKG] (b) [GBM]

Figure 1. The output of the algorithms for the network of 20 agents on MKG and GBM dataset. The performance of the distributed algorithms (ZGS-ESN, ZGS-FxTet-ESN and ZGS-FxTdt-ESN) is averaged across the agents.

(a) [MKG] (b) [GBM]

Figure 2. Evolution of testing error (NRMSE) for networks in the range of 4 to 20 agents.

(a) [MKG] (b) [GBM]

Figure 3. Evolution of training time for networks in the range of 4 to 20 agents.

Figure 4. Event-triggered instants of the agents.

NRMSE= n=1 N test ( y ˜ n y n ) 2 N test σ ^ t . (28)

In this context, σ ^ t signifies an empirical estimate of the variance in the actual output samples y 1 ,, y N test , where N test is the count of testing samples, and y ˜ 1 ,, y ˜ N test , denote the predicted outputs.

4.3. ESN Architecture

A regularization factor of C=10 and a reservoir size of D res =300 were selected, and these choices proved to be effective in all situations. In the reservoir, we employ tanh( ) nonlinearities, and for the output function, a scaled identity f( s )= α t s is utilized. In all instances, the parameter α t is set by searching within the set { 0.1,0.3,,0.9 } . The matrix β in res is populated with entries drawn from a uniform distribution within the range [ α i , α i ] . The parameter α i is explored within the same range as α t . The matrix β out res , which links the output to the reservoir, is initialized as a complete matrix with entries drawn from a uniform distribution within the range [ α f , α f ] . The parameter α f is searched within the same range as α t . α f =0 is allowed for the case in which no output feedback is required.

The reservoir matrix β res res is initialized using a uniform distribution in the range of [ 1,1 ] . Subsequently, around 55% of its connections are set to zero to foster sparsity. Lastly, the matrix is adjusted in scale to achieve a target spectral radius ρ , with ρ being explored within the same range as α t . Furthermore, uniform noise is introduced during the state update of the reservoir, where the noise is sampled from a uniform distribution within the interval [ 0, 10 3 ] . Additionally, the initial D=100 elements are excluded from each sequence. Table 1 displays the ultimate configurations obtained through the grid-search process.

Table 1. Description of the parameters.

Dataset

ρ

α i

α t

α f

MKG

0.9

0.3

0.5

0

4.4. Numerical Results

The test outputs of the four algorithms are presented in Figure 1, which demonstrate the prediction capability of the ESN. ZGS-FxTdt-ESN consistently follows the performance of the centralized solution across all scenarios. In Figure 1, the output value of ZGS-FxTdt-ESN basically coincides with the actual output value, indicating that ZGS-FxTdt-ESN maintains a relatively accurate output value. With the increasing number of agents, five algorithms show their test errors in Figure 2. We can see the test error of ZGS-FxTdt-ESN closely matches that of ZGS-FxTet-ESN. This shows that ZGS-FxTdt-ESN has the same test error as ZGS-FxTet-ESN, but ZGS-FxTdt-ESN has more advantages. Not only can its accuracy be comparable to that of ZGS-FxTet-ESN, but it also has a faster convergence speed. Figure 1 and Figure 2 indicate that the output values of C-ESN, ZGS-ESN and ZGS-FxTdt-ESN coincide and their test errors are all identical to that of ZGS-FxTet-ESN , which all suggests that their accuracies are similar. It has a faster convergence speed and only updates the state at the moment when the trigger condition is reached, which can save communication costs more. As shown in Figure 3, ZGS-FxTet-ESN achieves a significantly shorter training time than ZGS-FxTdt-ESN, primarily due to the latter’s need for continuous internal-state monitoring. In Figure 4, the event-triggered instants of agents are depicted, and it can be inferred that Zeno behavior is absent.

5. Conclusions

This paper presents a novel distributed learning algorithm designed for the training of a specialized recurrent neural network for the first time. The algorithm trained an identical ESN model among multiple agents. These agents were linked through a communication topology that operated without a central coordinator, relying solely on information exchange with neighboring nodes. The ZGS method was applied to solve the corresponding distributed optimization problem. The algorithm also incorporates the dynamic event-triggered method in order to reduce communication costs.

The effectiveness of the introduced algorithm was investigated by simulations. In the domain of distributed multi-agent systems, the convergence feature is a primary index for evaluating the performance of a distributed learning algorithm. Directions for future research could include considering dynamic handling of event triggering conditions to make it more applicable to real-life scenarios.

Acknowledgements

This work was supported in part by the National Natural Science Foundation of China (No. 62166013), the Natural Science Foundation of Guangxi (No. 2022GXNSFAA035499) and the Foundation of Guilin University of Technology (No. GLUTQD2007029).

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Zhang, H., Wang, Z. and Liu, D. (2014) A Comprehensive Review of Stability Analysis of Continuous-Time Recurrent Neural Networks. IEEE Transactions on Neural Networks and Learning Systems, 25, 1229-1262.
https://doi.org/10.1109/tnnls.2014.2317880
[2] Verykios, V.S., Bertino, E., Fovino, I.N., Provenza, L.P., Saygin, Y. and Theodoridis, Y. (2004) State-of-the-Art in Privacy Preserving Data Mining. ACM SIGMOD Record, 33, 50-57.
https://doi.org/10.1145/974121.974131
[3] Zhu, M., Zhou, J., Cai, H., He, X. and Xiao, F. (2023) A Distributed Learning Method for Deep Echo State Network Based on Improved SVD and ADMM. 2023 IEEE 20th International Conference on Mobile Ad Hoc and Smart Systems (MASS), Toronto, 25-27 September 2023, 270-278.
https://doi.org/10.1109/mass58611.2023.00040
[4] He, Y., Chen, L., Xu, Y., Zhu, Q. and Lu, S. (2023) A New Distributed Echo State Network Integrated with an Auto-Encoder for Dynamic Soft Sensing. IEEE Transactions on Instrumentation and Measurement, 72, 1-8.
https://doi.org/10.1109/tim.2022.3228278
[5] Alaviani, S.S. and Elia, N. (2019) Distributed Average Consensus over Random Networks. 2019 American Control Conference (ACC), Philadelphia, 10-12 July 2019, 1854-1859.
https://doi.org/10.23919/acc.2019.8815134
[6] Wang, H. and Yan, X. (2015) Optimizing the Echo State Network with a Binary Particle Swarm Optimization Algorithm. Knowledge-Based Systems, 86, 182-193.
https://doi.org/10.1016/j.knosys.2015.06.003
[7] Long, J., Zhang, S. and Li, C. (2020) Evolving Deep Echo State Networks for Intelligent Fault Diagnosis. IEEE Transactions on Industrial Informatics, 16, 4928-4937.
https://doi.org/10.1109/tii.2019.2938884
[8] Zhang, B., Han, Q. and Zhang, X. (2017) Recent Advances in Vibration Control of Offshore Platforms. Nonlinear Dynamics, 89, 755-771.
https://doi.org/10.1007/s11071-017-3503-4
[9] Lu, Y.M. and Do, M.N. (2008) A Theory for Sampling Signals from a Union of Subspaces. IEEE Transactions on Signal Processing, 56, 2334-2345.
https://doi.org/10.1109/tsp.2007.914346
[10] Zhang, X., Han, Q. and Yu, X. (2016) Survey on Recent Advances in Networked Control Systems. IEEE Transactions on Industrial Informatics, 12, 1740-1752.
https://doi.org/10.1109/tii.2015.2506545
[11] Heemels, W.P.M.H., Sandee, J.H. and Van Den Bosch, P.P.J. (2008) Analysis of Event-Driven Controllers for Linear Systems. International Journal of Control, 81, 571-590.
https://doi.org/10.1080/00207170701506919
[12] Yi, X., Yang, T., Wu, J. and Johansson, K.H. (2019) Distributed Event-Triggered Control for Global Consensus of Multi-Agent Systems with Input Saturation. Automatica, 100, 1-9.
https://doi.org/10.1016/j.automatica.2018.10.032
[13] Berneburg, J. and Nowzari, C. (2019) Distributed Dynamic Event-Triggered Coordination with a Designable Minimum Inter-Event Time. 2019 American Control Conference (ACC), Philadelphia, 10-12 July 2019, 1424-1429.
https://doi.org/10.23919/acc.2019.8815227
[14] Chen, J., Yue, D., Dou, C., Weng, S., Xie, X., Li, Y., et al. (2022) Static and Dynamic Event-Triggered Mechanisms for Distributed Secondary Control of Inverters in Low-Voltage Islanded Microgrids. IEEE Transactions on Cybernetics, 52, 6925-6938.
https://doi.org/10.1109/tcyb.2020.3034727
[15] Girard, A. (2015) Dynamic Triggering Mechanisms for Event-Triggered Control. IEEE Transactions on Automatic Control, 60, 1992-1997.
https://doi.org/10.1109/tac.2014.2366855
[16] Lu, J., Zhang, X., Zhang, B., Hou, X. and Wang, P. (2022) Distributed Dynamic Event-Triggered Control for Voltage Restoration and Current Sharing in DC Microgrids. IEEE Transactions on Sustainable Energy, 13, 619-628.
https://doi.org/10.1109/tste.2021.3123372
[17] Chen, J., Zhang, H. and Yin, G. (2023) Distributed Dynamic Event-Triggered Secure Model Predictive Control of Vehicle Platoon against Dos Attacks. IEEE Transactions on Vehicular Technology, 72, 2863-2877.
https://doi.org/10.1109/tvt.2022.3215966
[18] Ai, W. and Wang, D. (2020) Distributed Stochastic Configuration Networks with Cooperative Learning Paradigm. Information Sciences, 540, 1-16.
https://doi.org/10.1016/j.ins.2020.05.112
[19] Li, S., Ai, W. and Ge, X. (2020) Fixed-Time Distributed Cooperative Learning for Stochastic Configuration Networks. 2020 Chinese Automation Congress (CAC), Shanghai, 6-8 November 2020, 3380-3383.
https://doi.org/10.1109/cac51589.2020.9326656
[20] Hu, W., Yang, C., Huang, T. and Gui, W. (2020) A Distributed Dynamic Event-Triggered Control Approach to Consensus of Linear Multiagent Systems with Directed Networks. IEEE Transactions on Cybernetics, 50, 869-874.
https://doi.org/10.1109/tcyb.2018.2868778
[21] Scardapane, S., Wang, D. and Panella, M. (2016) A Decentralized Training Algorithm for Echo State Networks in Distributed Big Data Applications. Neural Networks, 78, 65-74.
https://doi.org/10.1016/j.neunet.2015.07.006
[22] Bapat, R.B. (2010) Graphs and Matrices, Volume 27. Springer.
[23] Zuo, Z. (2015) Nonsingular Fixed-Time Consensus Tracking for Second-Order Multi-Agent Networks. Automatica, 54, 305-309.
https://doi.org/10.1016/j.automatica.2015.01.021
[24] Fan, Y., Feng, G., Wang, Y. and Song, C. (2013) Distributed Event-Triggered Control of Multi-Agent Systems with Combinational Measurements. Automatica, 49, 671-675.
https://doi.org/10.1016/j.automatica.2012.11.010

Copyright © 2025 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.