Next Article in Journal
Research on Automatic Error Data Recognition Method for Structured Light System Based on Residual Neural Network
Next Article in Special Issue
PC-Allocation: Performance Cliff-Aware Two-Level Cache Resource Allocation Scheme for Storage System
Previous Article in Journal
Fermentation of Chenopodium formosanum Leaf Extract with Aspergillus oryzae Significantly Enhanced Its Physiological Activities
Previous Article in Special Issue
Quantum Image Steganography Schemes for Data Hiding: A Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Deep Echo State Network with Neuronal Similarity-Based Iterative Pruning Merging Algorithm

1
Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China
2
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(5), 2918; https://doi.org/10.3390/app13052918
Submission received: 22 January 2023 / Revised: 13 February 2023 / Accepted: 18 February 2023 / Published: 24 February 2023
(This article belongs to the Special Issue Advances in High-Performance Computing Research and Applications)

Abstract

:
Recently, a layer-stacked ESN model named deep echo state network (DeepESN) has been established. As an interactional model of a recurrent neural network and deep neural network, investigations of DeepESN are of significant importance in both areas. Optimizing the structure of neural networks remains a common task in artificial neural networks, and the question of how many neurons should be used in each layer of DeepESN must be stressed. In this paper, our aim is to solve the problem of choosing the optimized size of DeepESN. Inspired by the sensitive iterative pruning algorithm, a neuronal similarity-based iterative pruning merging algorithm (NS-IPMA) is proposed to iteratively prune or merge the most similar neurons in DeepESN. Two chaotic time series prediction tasks are applied to demonstrate the effectiveness of NS-IPMA. The results show that the DeepESN pruned by NS-IPMA outperforms the unpruned DeepESN with the same network size, and NS-IPMA is a feasible and superior approach to improving the generalization performance of DeepESN. The newly proposed method has broad application prospects in real-time systems.

1. Introduction

Recurrent neural networks (RNNs) represent a consolidated computational abstraction for learning with variable-length time series data [1]. As a simplified paradigm of RNN, the echo state network [2,3] (ESN) provides a prominent reduction in the computational cost compared to other paradigms of RNNs (e.g., [4] (LSTM), [5] (GRU)), for which the hidden layer of ESN is constructed by a randomly generated reservoir instead of independent neurons and the output weights of ESN are trained by a simple linear regression rather than a backpropagation algorithm. Thus, ESN has a successful application in various time series prediction problems (e.g., [6,7,8,9]).
Deep neural networks [10] (DNNs) have the potential to learn data representations at various levels of abstraction and are being increasingly stressed in the machine learning community. Recently, a layer-stacked ESN model named deep echo state network (DeepESN) was established and investigated, theoretically and experimentally, by Gallicchio et al: the inherent characterization of the system dynamics developed at the different layers of DeepESN is experimentally analyzed in [11] and theoretically explained in [12]; A theoretical foundation for the study of DeepESN from a dynamical system point of view is introduced in [13]; further details on the analysis and advancements of DeepESN could be found in [14]. As an interactional model of RNN and DNN, investigations of DeepESN are of significant importance in both areas. On the one hand, DeepESN expands our knowledge on how the information with memory attracted by RNN is extracted by hierarchical neural networks; on the other hand, DeepESN helps us better understand how the abstract intrinsic representations of time series extracted by DNN are recalled in reservoirs. Furthermore, DeepESN has richer nonlinear representation capacity, less computational complexity, and better predictive performance than single-layer ESN [15].
Optimizing the structure of neural networks remains a common task in artificial neural networks, and the same question of how many neurons should be used in each layer must be stressed in all types of neural networks. If the neurons are too few, the architecture does not satisfy the error demand by learning from the data, whereas if there are too many neurons, learning leads to the well-known overfitting problem [16]. As far as we know, minimal research has been carried out on optimizing the architecture of DeepESN. For research that has already been carried out on DeepESN, the same number is commonly assigned in each layer, which is acceptable but not optimal. In this paper, our aim is to solve the problem of choosing the optimized size of DeepESN, especially the number of neurons in different layers.
In 2014, a sensitive iterative pruning algorithm (SIPA) was proposed by Wang and Yan [17] to optimize the simple cycle reservoir network (SCRN). The algorithm was used to prune the least sensitive neurons one by one according to the sensitivity analysis, and the results showed that the SIPA method can optimize the structure and improve the generalization performance of the SCRN; meanwhile, pruning out redundant neurons could contribute to reducing the calculation and improving the computing efficiency of the network. Inspired by these advantages of SIPA, we wanted to apply a similar iterative pruning approach to DeepESN. However, SCRN is a kind of minimal-complexity ESN with a simple cycle topology, and the topology of DeepESN is much more complex than SCRN. Pruning a neuron in the network will raise perturbations on adjacent neurons, resulting in unstable network performance; in SIPA, the perturbations could be eliminated by adjusting the input weights into the perturbed neurons to minimize the distance of its input signal before and after pruning. Due to the hierarchical structure of reservoirs in DeepESN, perturbation raised by pruning a neuron in the lower layer will propagate into higher layers layer-by-layer, leading to greater instability of network performance and the difficulty of perturbation elimination.
In order to overcome the above difficulty, a new neuronal similarity-based iterative pruning merging algorithm (NS-IPMA) is proposed to iteratively prune out or merge the most similar neurons in DeepESN. In NS-IPMA, a pair of the most similar twin neurons, which are regarded as redundant neurons in the network, are selected out iteratively; then, if they exist in different layers, the one in the higher layer will be pruned out, and if they are in the same reservoir, they will be merged into one neuron, which works as the substitution of antecedent twin neurons. Quantitative estimation of neuronal similarity plays an essential role in determining the redundant neurons that should be pruned out. Four neuronal similarity estimation criteria of NS-IPMA approaches were attempted, including the inverse of Euclidean distance, Pearson’s correlation, Spearman’s correlation and Kendall’s correlation. Reducing the network size is a directly effective approach to improve the generalization performance of a neural network because pruning out neurons will lead to a reduction in network size to verify the effectiveness of the NS-IPMA method. The DeepESNs pruned by the NS-IPMA method were compared with unpruned DeepESNs, whose number of neurons in each reservoir is the same. The pruned DeepESN and the unpruned DeepESN were compared with an equal number of layers and equal number of total neurons. The results of the experiment on two chaotic time series prediction tasks showed that the NS-IPMA method has good network structure adaptability, and the DeepESNs pruned by the NS-IPMA method has better generalization performance and better robustness than the unpruned DeepESNs, indicating that the NS-IPMA method is a feasible and superior approach to improving the generalization performance of DeepESN. The NS-IPMA method provides a novel approach for choosing the appropriate network size of DeepESN, and it also has application potential in other RNNs and DNNs.
Our highlight contributions consist of:
  • A new NS-IPMA that works effectively in improving the generalization performance of DeepESN is proposed.
  • The proposed method contributes to simplifying the structure of DeepESN and reducing computational cost.
This paper is organized as follows: DeepESN and entropy quantification of reservoir richness are introduced in detail in Section 2, SIPA and the newly proposed NS-IPMA are described in Section 3, experiments and results are presented and discussed in Section 4 and Section 5, and Section 6 draws conclusions and highlights the future direction.

2. Deep Echo State Network

2.1. Leaky Integrator Echo State Network

A leaky integrator echo state network [18], (LI-ESN), as shown in Figure 1, is a recurrent neural network with three layers: input layer u ( t ) R N u × 1 , hidden layer x ( t ) R N x × 1 , output layer y ( t ) R N y × 1 . t denotes time sequence order. The hidden layer is regarded as a reservoir, which holds the memory of foregone information, and x ( t ) is refreshed by the state transition function:
x ( t ) = α x ( t 1 ) + ( 1 α ) tanh ( W i u ( t 1 ) + W r x ( t 1 ) ) ,
where W i R N x × N u is the input weight matrix randomly generated before training, and W r R N x × N x is the reservoir weight matrix previously given before training. α [ 0 , 1 ] is the leaky parameter. tanh ( ) is the activation function of the hidden layer. The reservoir weights in W r must be initialized to satisfy the echo state property (ESP) [19,20], denoting by ρ ( ) the spectral radius operator (i.e., the largest absolute eigenvalue of its matrix argument), and the necessary condition for the ESP is expressed as follows:
ρ ( 1 α ) I + α W r < 1 .
Accordingly, the values in matrix W r are randomly selected from a uniform distribution (e.g., U [ 1 , 1 ] ), and then rescaled to satisfy the above condition in Equation (2).
The output y ( t ) can be calculated through a linear combination of reservoir states as follows:
y ( t ) = W o x ( t ) ,
where W o R N y × N x is the output weight matrix.
During training, the states of reservoir neurons are collected in a training state matrix X t r a i n = [ x ( 1 ) , x ( 2 ) x ( L t r a i n ) ] , and an output target matrix Y t r a i n = [ y ^ ( 1 ) , y ^ ( 2 ) y ^ ( L t r a i n ) ] is collected correspondingly, where L t r a i n is the number of training samples.
The output weights in W o can be calculated by ridge regression as follows:
W o = ( X t r a i n T X t r a i n + λ I ) 1 X t r a i n T Y t r a i n T ,
where ( ) T represents matrix transpose, ( ) 1 represents matrix inversion, λ I is a regularization term ensuring X t r a i n T X t r a i n is invertible.

2.2. Deep Echo State Network

DeepESN, as shown in Figure 2, was first introduced by Gallicchio [11,14], as a stacked reservoir computing (RC) architecture, where multiple reservoir layers are stacked on top of each other. The state transition functions of hidden layers in DeepESN are expressed as:
x 1 ( t ) = α 1 x 1 ( t ) + ( 1 α 1 ) tanh ( W i u ( t 1 ) + W r 1 x 1 ( t 1 ) ) , l = 1 x l ( t ) = α l x l ( t ) + ( 1 α l ) tanh ( W p l 1 x l 1 ( t 1 ) + W r l x l ( t 1 ) ) , l [ 2 , L ]
where the superscript(1 and l) is the layer notation, with a total of L hidden layers in the network. x l ( t ) R N x l × 1 represents the l-th hidden layer (i.e., reservoir ( l ) ) with N x l neurons inside, W i R N x 1 × N u is the input weight matrix of the first hidden layer, W r l R N x l × N x l is the reservoir weight matrix of the l-th hidden layer, W p l 1 R N x l 1 × N x l is the propagate weight matrix, which connects reservoir ( l 1 ) to reservoir ( l ) .
As in the standard LI-ESN approach, the reservoir weights of a DeepESN are initialized subject to similar stability constraints. In the case of DeepESN, such constraints are expressed by the necessary conditions for the ESP of deep RC networks [13] described by the following equation:
max l [ 1 , L ] ρ ( 1 α l ) I + α l W r < 1 ,
where the same leaky parameter ( α l α ) in each layer is considered in this paper.
The values in each reservoir matrix { W r l | l [ 1 , L ] } are randomly initialized from a uniform distribution (e.g., U [ 1 , 1 ] ); after that, each W r l is spectrally normalized by its spectral radius and rescaled by the same reservoir scaling parameter γ r to meet the demands of Equation (6).
W r l γ r W r l ρ ( W r l ) , l [ 1 , L ] .
The values in input weight matrix W i are randomly selected from a uniform distribution U [ γ i , γ i ] , where γ i is the input scaling parameter. The values in each propagate weight matrix { W p l | l [ 1 , L 1 ] } are randomly selected from a uniform distribution U [ γ p , γ p ] , where γ p is the propagate scaling parameter.
The output equation and the training equation of DeepESN are formed by concatenating all hidden neurons in all reservoirs together, denoting x ˜ ( t ) T = [ x 1 ( t ) T x 2 ( t ) T x L ( t ) T ] and substituting x ˜ ( t ) for x ( t ) in Equations (3) and (4):
y ( t ) = W o x ˜ ( t ) ,
W o = ( X ˜ t r a i n T X ˜ t r a i n + λ I ) 1 X ˜ t r a i n T Y t r a i n T .

2.3. Architectural Richness of DeepESN

The components of the state should be as diverse as possible to provide a richer pool of dynamics from which the trainable part can appropriately combine. From an information-theoretic point of view, this form of richness could be measured by means of the entropy of instantaneous reservoir states [1]. Here, an efficient estimator of Renyi’s quadratic entropy is introduced: suppose that we have N independent and identically distributed samples { v 1 , , v N } for the continuous random variable V . An estimation of Renyi’s quadratic entropy directly from the sampling data is defined as:
H 2 ( V ) = log ( 1 N 2 i = 1 N j = 1 N G κ 2 ( v j v i ) ) ,
where G κ 2 ( ) is a Gauss kernel function with standard deviation κ 2 , and κ can be determined by Silverman’s rule:
κ = σ ( V ) ( 4 N 1 ( 2 d + 1 ) 1 ) 1 d + 4 ,
where σ ( V ) is the standard deviation, and d is the data dimensionality.
Average state entropy (ASE) is obtained by time averaging the instantaneous Renyi’s quadratic entropy estimation of reservoir neurons.
H ( t ) = H 2 ( x ˜ ( t ) ) ,
A S E = 1 S t = 1 S H ( t ) ,
where S is the sample number. ASE gives us a research perspective independent of the learning aspect, and higher ASE values are preferable and denote richer dynamics in reservoirs [1].

3. Pruning Deep Echo State Network with a Neuronal Similarity-Based Iterative Pruning Merging Algorithm

3.1. Sensitive Iterative Pruning Algorithm to Simple Cycle Reservoir Network

The SCRN is a kind of minimum complexity ESN, which has a cycle topology in the reservoir [21]. Every reservoir neuron is unidirectionally connected to its two adjacent neurons. An SIPA method is introduced to choose the right network size for SCRN by iteratively pruning out the least sensitive neurons [17].
The SIPA method is carried out in the following steps:
  • Establish an SCRN with a large enough reservoir and satisfactory performance.
  • Select a neuron to be pruned using the sensitive criterion [22] (assume x m is to be pruned), as diagramed in Figure 3, and remove all weights connected to x m as follows:
    W i ( m , : ) 0 W r ( m , m 1 ) 0 W r ( m + 1 , m ) 0
  • Establish a new link between two neighbors of pruned neurons, the link weight is determined to eliminate the perturbation caused by pruning, denoting the input to x m + 1 before pruning I o = W i ( m + 1 , : ) u + W r ( m + 1 , m ) x m , the input to x m + 1 after pruning I n = W i ( m + 1 , : ) u + W r ( m + 1 , m 1 ) x m 1 , the perturbation is eliminated, and the original reservoir behavior is maintained as long as I n is set as close as possible to I o , by solving the following optimization problem:
    min W r ( m + 1 , m 1 ) | | I o I n | | 2 .
  • Adjust the output weights by retraining the network, and then calculate the training error.
  • Repeat steps 2–4 until the training error or the reservoir size reaches an acceptable range.
The key to a successful application of SIPA is Step 3. In Step 2, a reservoir perturbation is triggered by pruning, making the performance of the pruned network unpredictable. Thus, the essential task of Step 3 is to reduce the effects of perturbation so that the rest of the neurons remain unchanged and the network performance is approximately as good as before.

3.2. Neuronal Similarity-Based Iterative Pruning Merging Algorithm to Deep Echo State Network

Due to the simple topology of SCRN, only one neuron ( x m + 1 ) receives input from the pruned neuron ( x m ), and perturbation elimination in SCRN is easy to perform. However, the topology of DeepESN is much more complicated, and more extensive perturbation will be raised by pruning one neuron in DeepESN because, in a highly coupled reservoir, perturbation at any neuron will be diffused to every neuron of the same reservoir. Furthermore, in hierarchically stacked reservoirs, perturbation in the lower layer will be transmitted to every higher layer above, and it is very difficult to eliminate all these perturbations.
The merging method is designed to solve this difficulty, and this coincides with Islam’s idea [23], where two neurons are merged by averaging their input weights. Consider the following ideal scenario: merging two identical twin neurons in the same reservoir will derive a new neuron that is identical to two antecedent twins; this new burned neuron could act as an equivalent substitution for two antecedent twins. Consequently, a neuron is pruned without leading any perturbation through superimposing the output weights of the merged twins as well. This ideal perturbation-free merging has the prerequisite that two identical neurons have to be found in the same reservoir. The more similar the two merged neurons are, the weaker perturbation that will be raised by merging. Neuronal similarity could be assessed by some quantitative relations of the collected training state matrix. Distance and correlation are commonly used to quantify similarity, and four similarity estimation criteria are given in this paper, including the inverse of the Euclidean distance ( E D ), Pearson’s correlation coefficient ( P C ), Spearman’s correlation coefficient ( S C ) and Kendall’s correlation coefficient ( K C ) as follows:
Noting the total number of neurons in all reservoirs M = l = 0 L N x l , recall the train state matrix
X ˜ t r a i n = x 1 ( 1 ) x 1 ( 2 ) x 1 ( L t r a i n ) x 2 ( 1 ) x 2 ( 2 ) x 1 ( L t r a i n ) x L ( 1 ) x L ( 2 ) x L ( L t r a i n ) ( M × L t r a i n ) = n 1 n 2 n M ,
where L t r a i n is the number of training samples, and n i is the historical state of the i-th neuron state during training. The similarity of n i and n j is derived by:
E D ( i , j ) = 1 n i n j 2 ,
P C ( i , j ) = σ n i n j σ n i n i σ n j n j ,
S C ( i , j ) = 1 6 t = 1 L t r a i n d t 2 L t r a i n L t r a i n 2 1 ,
K C ( i , j ) = c d 1 2 L t r a i n L t r a i n 1 ,
where σ n i n j represents the cross-correlation of n i and n j , and σ n i n i and σ n j n j represent the autocorrelation of n i and n j . d t is the rank difference of n i ( t ) and n j ( t ) , c is the number of concordant pairs, and d is the number of discordant pairs in n i and n j . NS-IPMA based on different similarity estimation criteria are named correspondingly; for instance, ES-IPMA means NS-IPMA based on the inverse of Euclidean distance criterion, etc. The main process of NS-IPMA is illustrated in Figure 4, and the NS-IPMA method is carried out in the following steps:
  • Initially generate a performable DeepESN with large enough reservoirs by tuning hyperparameters to minimize the average of training and validate error using the particle swarm optimization (PSO) algorithm; please refer to Appendix A for more details on hyperparameter tuning. This DeepESN is a primitive network to implement.
  • Washout the reservoirs, and activate the reservoirs using training samples to obtain the training state matrix.
  • Quantify the similarity of every two neurons (using one criterion of Equations (17)–(20)), and select a pair of the most similar neurons.
  • (a)
    If selected neurons are in the same reservoir (note as x i l and x j l ), merge them. As diagramed in Figure 5, x s l is the son neurons merged as the substitute of its parents x i l and x j l . l is the reservoir layer where x i l and x j l exist. The related weight matrix W p l 1 , W p l and W r l is refreshed as follows:
    W p l 1 ( s , : ) W p l 1 ( i , : ) + W p l 1 ( j , : ) 2 W r l ( s , : ) W r l ( i , : ) + W r l ( j , : ) 2 W r l ( : , s ) W r l ( : , i ) + W r l ( : , j ) W p l ( : , s ) W p l ( : , i ) + W p l ( : , j )
    if l = 1 , W i perform as W p l 1 ; if l = L , W p l does not exist.
    (b)
    If selected neurons are in different reservoirs (note as x i m and x j l ), prune one in a high layer (assume m < l ). The related weight matrix W p l 1 , W p l and W r l is refreshed as follows:
    W p l 1 ( j , : ) 0 W r l ( j , : ) 0 W r l ( : , j ) 0 W p l ( : , j ) 0
    if l = 1 , W i perform as W p l 1 ; if l = L , W p l does not exist.
  • Adjust the output weights by retraining the network, and then estimate the performance of the current network.
  • Repeat steps 2–5 until the training error or the network size reaches an acceptable range.

4. Experiments

Due to high nonlinearity and instability, forecasting chaotic time series has remained a difficult task for the last few decades [24]. RC architectures, including ESNs, have successfully been employed for multi-step-ahead prediction of nonlinear time series and modeling chaotic dynamical systems at a low computational cost [25]; therefore, chaotic time series prediction is an ideal task to estimate the performance of DeepESN.

4.1. Datasets

4.1.1. Mackey–Glass Chaotic Time Series

The Mackey–Glass time series [26] is a standard benchmark for chaotic time series forecast models. The Mackey-Glass time series is defined in the following differential equation:
d x d t = a x τ 1 + x τ c b x , x τ = x ( t τ ) ,
where a = 0.1 , b = 0.2 , and c = 11 are constant parameters. The nonlinearity of the system increases as the time delay parameter τ increases. The system shows chaotic behavior when τ 17 . To generate the time series used here, τ is set to 31, and the integration step size is set to Δ t = 0.002   s using the jitcdde Python engine to solve delay differential equations at discrete equally spaced times. The data were then sampled by 4 Δ t to form the Mackey–Glass ( M G ) dataset with 4000 data points. A Python library [27] was used to implement this process.

4.1.2. Lorenz Chaotic Time Series

The second chaotic time series benchmark is derived from the Lorenz system [28], which is given by the following equations:
d x d t = a ( x y ) , d y d t = b ( x y ) + x z , d z d t = c z + x y ,
where a = 34 3 , b = 298 11 , and c = 17 7 are the constant parameters. The time series is obtained by numerical integration of the equation using odeint, a Python solver of the ordinary differential equation system, where the solution is evaluated at times spaced Δ t = 0.08 s apart, and the initial values are set to [ x ( 0 ) , y ( 0 ) , z ( 0 ) ] = [ 0.5 , 0 , 0.5 ] , and the Lorenz z-axis ( L Z ) dataset is obtained by taking the z-axis data from 4000 samples.
Before carrying out the prediction task, both datasets are shifted by their mean to remove the DC bias.

4.2. Next Spot Prediction Task

In the next spot prediction experiment, the last four continuous spots were considered the input and the next spot was considered the desired output, i.e., { u ( t 3 ) , u ( t 2 ) , u ( t 1 ) , u ( t ) } was used to predict u ( t + 1 ) . The number of the train samples and the test samples was 1800 and 2000, 200 samples before the first training sample was used to wash out the initial transient (see Figure 6). The reservoir diversity was quantized by the ASE (see: Section 2.3) of the combined training and testing neuron states activated by training and testing samples. The predictive error performance was quantized by the normalized root mean square error:
N R M S E = t = 1 T y ( t ) y ^ ( t ) 2 T y ( t ) y ¯ ( t ) 2 ,
where T is sample number, y ^ ( t ) is the desired output, y ( t ) is the readout output, and y ¯ ( t ) is the average of y ( t ) .
Two different initial reservoir size conditions of DeepESN were performed on both datasets. The first is four stacked reservoirs with 100 neurons in each reservoir (abbreviated to: 4 × 100 ), and the second is eight stacked reservoirs with 50 neurons in each reservoir (abbreviated to: 8 × 50 ). All experiments were carried out under two different initial conditions on two datasets, and model hyperparameters tuned by PSO were recorded in Table A1.

4.3. Ablation Experiment and Control Experiment

In order to demonstrate the effectiveness of different similarity estimation criteria, the worst-case scenario of NS-IPMA, the iterative pruning merging algorithm (IPMA) without neuronal similarity estimation, was investigated as an ablation experiment. In IPMA, the similarity of each pair of two neurons was assigned by random values; thus, two random neurons would be recognized as the most similar pair of neurons and would be pruned (or merged).
To verify the effectiveness of the NS-IPMA method, the pruned DeepESNs were compared with a control experiment. The unpruned DeepESN, whose number of neurons in each reservoir is the same, the pruned DeepESN, and the unpruned DeepESN were compared with an equal number of layers, equal number of total neurons and the same hyperparameters because the unpruned DeepESN is a standard benchmark that identifies the evolutionary characteristic of network performance by reducing network size. Ninety percent of neurons of randomly initialized DeepESNs were continuously pruned using different criterion-based NS-IPMA methods (ED-IPMA, PC-IPMA, SC-IPMA, KC-IPMA) and non-criterion-based IPMA. During pruning, networks were silhouetted, and the performance was evaluated once 10% of neurons had been pruned. These pruned groups were compared with unpruned DeepESN. All experiments were repeated 20 times, and all results were averaged through 20 independent replications.

5. Results and Discussion

5.1. Hierarchical Structure

The number of neurons remaining during pruning in each reservoir of DeepESN, which was pruned by different similarity estimation criterion-based NS-IPMA methods, is shown in Figure 7. As NS-IPMA goes on, we observed a significant reduction in the neuron number in high layers at a later stage. The reason for this was the discard policy, in which when a redundant pair of the most similar neurons were found in different reservoirs, the one in the higher layer would be pruned out. Lower layer reservoirs are the foundation of higher layer reservoirs, and too few neurons in lower layers bring the risk of insufficient information presentation and extraction in higher layers; thus, the NS-IPMA methods have good network structure adaptability.

5.2. Reservoir Diversity and Error Performance

Quantitative performance comparisons (ASE, training NRMSE, and testing NRMSE) of unpruned DeepESN and DeepESN, which were pruned by IPMA and different similarity estimation criterion-based NS-IPMA methods, are illustrated in Figure 8, Figure 9 and Figure 10. The results on the two datasets were similar but different. On the whole, as reservoir size reduced, the ASE would decrease and training NRMSE would increase, confirming that the model with more neurons has a better information representation ability and a better training effect. Initially, testing NRMSE was greater than training NRMSE; this phenomenon is called overfitting. As the reservoir size reduced, the training error went down, and the overfitting was improved, confirming that reducing the network size is a directly effective approach to improving the generalization performance of the network. There is no doubt that pruning neurons will lead to a reduction in network size, which is the reason why the unpruned DeepESN is tested as a benchmark, and a successful pruning algorithm should outperform this benchmark. Furthermore, the minimum testing NRMSE condition is chosen to compare the extreme generalization performance of each experiment group, which was recorded in Table 1.
Similarly, from Figure 8, the DeepESN pruned by ED-IPMA maintained a minimum ASE loss as the number of neurons decreases, and we suspect that there were some hidden relationships between the Euler distance and Renyi’s quadratic entropy. From Table 1, the unpruned DeepESNs had a greater standard deviation of the minimum testing NRMSE compared to the pruned DeepESN, indicating that the NS-IPMA method has good robustness.
On the MG dataset: From Figure 10a,b, in the early stage, the testing error of DeepESN pruned by SC-IPMA and KC-IPMA dropped rapidly; later, DeepESN pruned by ED-IPNA and PC-IPMA performed best when the majority of neurons had been pruned. From Table 1, although ED-IPMA achieved the best extreme generalization performance, the pruned DeepESNs had no obvious improvement in the mean value of minimum testing NRMSE compared to the unpruned DeepESN.
On the LZ dataset: From Figure 9c,d, Figure 10c,d and Table 1, the training error, testing error, and extreme generalization performance of pruned DeepESNs were all significantly improved compared to unpruned DeepESN, and the diversity of different similarity estimation criteria were not prominent. The noncriterion-based IPMA method performed as well as criterion-based NS-IPMA methods.
In summary, all these experimental results showed that the NS-IPMA method is a successful approach to improving the generalization performance of DeepESN, which is specific, in almost all of our designed experiments, in that the DeepESN pruned by NS-IPMA has better generalization performance than the standard unpruned DeepESN.

6. Conclusions and Prospects

In our research, a new iterative pruning merging algorithm was proposed to simplify the architecture of DeepESN. As to which neurons should be pruned out, four different similarity estimation criteria were attempted. The unpruned DeepESNs is a benchmark that identifies the evolutionary characteristic of network performance by reducing network size, and the effectiveness of the proposed method was experimentally verified by comparing pruned DeepESNs with unpruned DeepESNs in the same network size. The results showed that these NS-IPMA methods have good network structure adaptability, and the DeepESNs pruned by the NS-IPMA method have better generalization performance and better robustness than unpruned DeepESNs, indicating that the NS-IPMA method is a feasible and superior approach to improving the generalization performance of DeepESN. The NS-IPMA method provides a novel approach for choosing the appropriate network size of DeepESN. One could start with a larger model than necessary for reservoir size and then prune or merge some similar neurons to obtain a better DeepESN model. One could select a simple architecture with small computation requirements while keeping the testing error acceptable.
In many tasks, low computational cost and reliable performance cannot be achieved at the same time. The experimental results showed that the newly proposed method could better balance the computational cost and performance; therefore, it has broad application prospects in real-time (RT) systems, including RT control, RT forecast, RT decision, etc., for example, RT control for modernized microgrids [29], intelligent Internet of Things automatic irrigation control system [30], lower generation data prediction for wind power plant [31], trading prediction and dynamic decisions for online trading systems [32], model predictive control [33,34], predictive feedforward control for servo systems [35,36].
However, there are still some shortcomings in our work. First, the problem of how to choose the redundant neurons to be pruned out or what the best neuronal similarity estimation criterion should be, remains unsolved. Second, only the hierarchical structure, the reservoir diversity, and the overall error performance are investigated, and more evolutionary characteristics of different reservoirs, such as their spectral radius, resulting from the NS-IPMA method, are not analyzed. Third, the effects of pruning and merging are not clearly distinguished. Future research on the NS-IPMA method could focus on the theoretical analysis of the above-unsolved problem or on practical applications where high-performance computing is required.

Author Contributions

Methodology, Q.S.; software, Q.S.; validation, Q.S.; investigation, Q.S.; resources, Q.S.; writing—original draft, Q.S.; writing—review and editing, Q.S. and H.Z.; supervision, Y.M.; funding acquisition, Y.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China—grant number: 62271109.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this research and relevant resources are available upon request; please send your request message to shen99855@outlook.com.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RNNRecurrent neural network
ESNEcho state network
LSTMLong short-term memory
GRUGated recurrent unit
DNNDeep neural network
DeepESNDeep echo state network
SIPASensitive iterative pruning algorithm
SCRNSimple cycle reservoir network
NS-IPMANeuronal similarity-based iterative pruning merging algorithm
LI-ESNLeaky integrator echo state network
ESPEcho state property
RCReservoir computing
ASEAverage state entropy
EDEuclidean distance
PCPearson’s correlation
SCSpearman’s correlation
KCKendall’s correlation
ED-IPMANS-IPMA based on the inverse of Euclidean distance criterion
PC-IPMANS-IPMA based on the inverse of Pearson’s correlation criterion
SC-IPMANS-IPMA based on the inverse of Spearman’s correlation criterion
KC-IPMANS-IPMA based on the inverse of Kendall’s correlation criterion
IPMAIterative pruning merging algorithm
PSOParticle swarm optimization
M G Mackey–Glass
L Z Lorenz z-axis
DCDirect component
NRMSENormalized root mean square error
RTReal-time

Appendix A. Hyperparameter Tuning

As defined in Section 2.2, hyperparameters ( α , γ i , γ r , γ p and λ ) play essential roles in the performance of DeepESN, as well as in the successful application of NS-IPMA. α and γ r affect the stability of reservoirs, α = 0.92 and γ r = 0.8 are set to satisfy Equation (6); λ affects the generalization performance, a small regularization factor λ = 1 × 10 10 is chosen to make the output weights better fit the training samples. γ i adjusts the strength of the input signal into the first layer of DeepESN, γ p adjusts the strength of the input signal into higher layers of DeepESN; thus, γ i is optimized on the LI-ESN, which is the first layer of DeepESN. After that, γ p is optimized on the DeepESN after higher layers are hierarchically stacked on the original LI-ESN. The tuned results are recorded in Table A1.
Table A1. Hyperparameters applied under different initial reservoir sizes on different datasets.
Table A1. Hyperparameters applied under different initial reservoir sizes on different datasets.
Dataset MG MG LZ LZ
Initial Size 4 × 100 8 × 50 4 × 100 8 × 50
α 0.920.920.920.92
γ r 0.80.80.80.8
γ i * 1 0.373 840.253 140.096 310.064 94
γ p * 2 0.211 360.241 620.335 510.347 11
λ 1 × 10 10 1 × 10 10 1 × 10 10 1 × 10 10
*1 Tuned by the PSO in the range of [1 × 10−5, 10]; *2 tuned by PSO in the range of [0.1, 5].

References

  1. Gallicchio, C.; Micheli, A. Architectural richness in deep reservoir computing. Neural Comput. Appl. 2022, 8. [Google Scholar] [CrossRef]
  2. Jaeger, H.; Haas, H. Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication. Science 2004, 304, 78–80. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Zhang, S.; He, K.; Cabrera, D.; Li, C.; Bai, Y.; Long, J. Transmission Condition Monitoring of 3D Printers Based on the Echo State Network. Appl. Sci. 2019, 9, 3058. [Google Scholar] [CrossRef] [Green Version]
  4. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  5. Che, Z.P.; Purushotham, S.; Cho, K.; Sontag, D.; Liu, Y. Recurrent Neural Networks for Multivariate Time Series with Missing Values. Sci. Rep. 2018, 8, 6085. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Xu, X.; Ren, W. Prediction of Air Pollution Concentration Based on mRMR and Echo State Network. Appl. Sci. 2019, 9, 1811. [Google Scholar] [CrossRef] [Green Version]
  7. Zhang, M.; Wang, B.; Zhou, Y.; Sun, H. WOA-Based Echo State Network for Chaotic Time Series Prediction. J. Korean Phys. Soc. 2020, 76, 384–391. [Google Scholar] [CrossRef]
  8. Zhang, M.; Wang, B.; Zhou, Y.; Gu, J.; Wu, Y. Prediction of Chaotic Time Series Based on SALR Model with Its Application on Heating Load Prediction. Arab. J. Sci. Eng. 2021, 46, 8171–8187. [Google Scholar] [CrossRef]
  9. Zhou, J.; Wang, H.; Xiao, F.; Yan, X.; Sun, L. Network traffic prediction method based on echo state network with adaptive reservoir. Softw. Pract. Exp. 2021, 51, 2238–2251. [Google Scholar] [CrossRef]
  10. Baek, J.; Choi, Y. Deep Neural Network for Predicting Ore Production by Truck-Haulage Systems in Open-Pit Mines. Appl. Sci. 2020, 10, 1657. [Google Scholar] [CrossRef] [Green Version]
  11. Gallicchio, C.; Micheli, A.; Pedrelli, L. Deep reservoir computing: A critical experimental analysis. Neurocomputing 2017, 268, 87–99. [Google Scholar] [CrossRef]
  12. Gallicchio, C.; Micheli, A.; Silvestri, L. Local Lyapunov exponents of deep echo state networks. Neurocomputing 2018, 298, 34–45. [Google Scholar] [CrossRef]
  13. Gallicchio, C.; Micheli, A. Echo State Property of Deep Reservoir Computing Networks. Cogn. Comput. 2017, 9, 337–350. [Google Scholar] [CrossRef]
  14. Gallicchio, C.; Micheli, A. Deep Echo State Network (DeepESN): A Brief Survey. arXiv 2017, arXiv:abs/1712.04323. [Google Scholar]
  15. Gallicchio, C.; Micheli, A. Why Layering in Recurrent Neural Networks? In A DeepESN Survey. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar] [CrossRef]
  16. Thomas, P.; Suhner, M.C. A New Multilayer Perceptron Pruning Algorithm for Classification and Regression Applications. Neural Process. Lett. 2015, 42, 437–458. [Google Scholar] [CrossRef]
  17. Wang, H.; Yan, X. Improved simple deterministically constructed Cycle Reservoir Network with Sensitive Iterative Pruning Algorithm. Neurocomputing 2014, 145, 353–362. [Google Scholar] [CrossRef]
  18. Jaeger, H.; Lukoševičius, M.; Popovici, D.; Siewert, U. Optimization and applications of echo state networks with leaky- integrator neurons. Neural Netw. 2007, 20, 335–352. [Google Scholar] [CrossRef]
  19. Yildiz, I.B.; Jaeger, H.; Kiebel, S.J. Re-visiting the echo state property. Neural Netw. 2012, 35, 1–9. [Google Scholar] [CrossRef]
  20. Gallicchio, C.; Micheli, A. Architectural and Markovian factors of echo state networks. Neural Netw. 2011, 24, 440–456. [Google Scholar] [CrossRef]
  21. Rodan, A.; Tino, P. Minimum complexity echo state network. IEEE Trans. Neural Netw. 2011, 22, 131–144. [Google Scholar] [CrossRef]
  22. Castellano, G.; Fanelli, A.; Pelillo, M. An iterative pruning algorithm for feedforward neural networks. IEEE Trans. Neural Netw. 1997, 8, 519–531. [Google Scholar] [CrossRef]
  23. Islam, M.M.; Sattar, M.A.; Amin, M.F.; Yao, X.; Murase, K. A New Adaptive Merging and Growing Algorithm for Designing Artificial Neural Networks. IEEE Trans. Syst. Man Cybern. Part (Cybern.) 2009, 39, 705–722. [Google Scholar] [CrossRef]
  24. Shahi, S.; Fenton, F.H.; Cherry, E.M. Prediction of chaotic time series using recurrent neural networks and reservoir computing techniques: A comparative study. Mach. Learn. Appl. 2022, 8, 100300. [Google Scholar] [CrossRef] [PubMed]
  25. Bianchi, F.M.; Maiorino, E.; Kampffmeyer, M.C.; Rizzi, A.; Jenssen, R. Other Recurrent Neural Networks Models. In Recurrent Neural Networks for Short-Term Load Forecasting: An Overview and Comparative Analysis; Springer International Publishing: Cham, Switzerland, 2017; pp. 31–39. [Google Scholar] [CrossRef]
  26. Mackey, M.C.; Glass, L. Oscillation and chaos in physiological control systems. Science 1977, 197, 287–289. [Google Scholar] [CrossRef]
  27. Maat, J.R.; Malali, A.; Protopapas, P. TimeSynth: A Multipurpose Library for Synthetic Time Series in Python. 2017. Available online: https://github.com/TimeSynth/TimeSynth (accessed on 22 February 2023).
  28. Chao, K.H.; Chang, L.Y.; Xu, F.Q. Smart Fault-Tolerant Control System Based on Chaos Theory and Extension Theory for Locating Faults in a Three-Level T-Type Inverter. Appl. Sci. 2019, 9, 3071. [Google Scholar] [CrossRef] [Green Version]
  29. Yan, S.R.; Guo, W.; Mohammadzadeh, A.; Rathinasamy, S. Optimal deep learning control for modernized microgrids. Appl. Intell. 2022. [Google Scholar] [CrossRef]
  30. Saini, S.S.; Soni, D.; Malhi, S.S.; Tiwari, V.; Goyal, A. Automatic irrigation control system using Internet of Things(IoT). J. Discret. Math. Sci. Cryptogr. 2022, 25, 879–889. [Google Scholar] [CrossRef]
  31. Ozkan, M.B.; Karagoz, P. Data Mining-Based Upscaling Approach for Regional Wind Power Forecasting: Regional Statistical Hybrid Wind Power Forecast Technique (RegionalSHWIP). IEEE Access 2019, 7, 171790–171800. [Google Scholar] [CrossRef]
  32. Liu, Z.; Luo, H.; Chen, P.; Xia, Q.; Gan, Z.; Shan, W. An efficient isomorphic CNN-based prediction and decision framework for financial time series. Intell. Data Anal. 2022, 26, 893–909. [Google Scholar] [CrossRef]
  33. Zhang, B.; Sun, X.; Liu, S.; Deng, X. Tracking control of multiple unmanned aerial vehicles incorporating disturbance observer and model predictive approach. Trans. Inst. Meas. Control 2020, 42, 951–964. [Google Scholar] [CrossRef]
  34. Xing, X.; Lin, J.; Wan, C.; Song, Y. Model Predictive Control of LPC-Looped Active Distribution Network With High Penetration of Distributed Generation. IEEE Trans. Sustain. Energy 2017, 8, 1051–1063. [Google Scholar] [CrossRef]
  35. Benrabah, M.; Kara, K.; AitSahed, O.; Hadjili, M.L. Constrained Nonlinear Predictive Control Using Neural Networks and Teaching-Learning-Based Optimization. J. Control. Autom. Electr. Syst. 2021, 32, 1228–1243. [Google Scholar] [CrossRef]
  36. Wang, Y.; Yu, H.; Che, Z.; Wang, Y.; Zeng, C. Extended State Observer-Based Predictive Speed Control for Permanent Magnet Linear Synchronous Motor. Processes 2019, 7, 618. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Structure of LI-ESN.
Figure 1. Structure of LI-ESN.
Applsci 13 02918 g001
Figure 2. Structure of DeepESN.
Figure 2. Structure of DeepESN.
Applsci 13 02918 g002
Figure 3. Weight coefficient diagram of connections of neurons in the reservoir of SCRN before (a) and after (b) pruning a neuron by SIPA.
Figure 3. Weight coefficient diagram of connections of neurons in the reservoir of SCRN before (a) and after (b) pruning a neuron by SIPA.
Applsci 13 02918 g003
Figure 4. Flowchart of NS-IPMA.
Figure 4. Flowchart of NS-IPMA.
Applsci 13 02918 g004
Figure 5. Weight coefficients diagram of connected neurons in reservoir (l) of DeepESN before (a) and after (b) merging a neuron by NS-IPMA.
Figure 5. Weight coefficients diagram of connected neurons in reservoir (l) of DeepESN before (a) and after (b) merging a neuron by NS-IPMA.
Applsci 13 02918 g005
Figure 6. M G (a) and L Z (b) datasets.
Figure 6. M G (a) and L Z (b) datasets.
Applsci 13 02918 g006
Figure 7. The figure shows the number of neurons remaining in each layer of DeepESN, pruned by different similarity estimation criterion-based NS-IPMA methods. The vertical axis indicates the layer index, the horizontal axis indicates the percent of pruned neurons in the initial total number of neurons, the mesh color indicates the number of remaining neurons, and the darker color indicates more neurons remaining. (a): Initial 4 layer reservoirs with 100 neurons in each reservoir on M G dataset; (b): Initial 8 layer reservoirs with 50 neurons in each reservoir on M G dataset; (c): Initial 4 layer reservoirs with 100 neurons in each reservoir on L Z dataset; (d): Initial 8 layer reservoirs with 50 neurons in each reservoir on L Z dataset.
Figure 7. The figure shows the number of neurons remaining in each layer of DeepESN, pruned by different similarity estimation criterion-based NS-IPMA methods. The vertical axis indicates the layer index, the horizontal axis indicates the percent of pruned neurons in the initial total number of neurons, the mesh color indicates the number of remaining neurons, and the darker color indicates more neurons remaining. (a): Initial 4 layer reservoirs with 100 neurons in each reservoir on M G dataset; (b): Initial 8 layer reservoirs with 50 neurons in each reservoir on M G dataset; (c): Initial 4 layer reservoirs with 100 neurons in each reservoir on L Z dataset; (d): Initial 8 layer reservoirs with 50 neurons in each reservoir on L Z dataset.
Applsci 13 02918 g007aApplsci 13 02918 g007b
Figure 8. ASE comparison of unpruned DeepESN and DeepESN pruned by IPMA, ED-IPMA, PC-IPMA, SC-IPMA, KC-IPMA. (a): Initial 4 layer reservoirs with 100 neurons in each reservoir on M G dataset; (b): Initial 8 layer reservoirs with 50 neurons in each reservoir on M G dataset; (c): Initial 4 layer reservoirs with 100 neurons in each reservoir on L Z dataset; (d): Initial 8 layer reservoirs with 50 neurons in each reservoir on L Z dataset.
Figure 8. ASE comparison of unpruned DeepESN and DeepESN pruned by IPMA, ED-IPMA, PC-IPMA, SC-IPMA, KC-IPMA. (a): Initial 4 layer reservoirs with 100 neurons in each reservoir on M G dataset; (b): Initial 8 layer reservoirs with 50 neurons in each reservoir on M G dataset; (c): Initial 4 layer reservoirs with 100 neurons in each reservoir on L Z dataset; (d): Initial 8 layer reservoirs with 50 neurons in each reservoir on L Z dataset.
Applsci 13 02918 g008
Figure 9. Training NRMSE comparison of unpruned DeepESN and DeepESN pruned by IPMA, ED-IPMA, PC-IPMA, SC-IPMA, KC-IPMA. (a): Initial 4 layer reservoirs with 100 neurons in each reservoir on M G dataset; (b): Initial 8 layer reservoirs with 50 neurons in each reservoir on M G dataset; (c): Initial 4 layer reservoirs with 100 neurons in each reservoir on L Z dataset; (d): Initial 8 layer reservoirs with 50 neurons in each reservoir on L Z dataset.
Figure 9. Training NRMSE comparison of unpruned DeepESN and DeepESN pruned by IPMA, ED-IPMA, PC-IPMA, SC-IPMA, KC-IPMA. (a): Initial 4 layer reservoirs with 100 neurons in each reservoir on M G dataset; (b): Initial 8 layer reservoirs with 50 neurons in each reservoir on M G dataset; (c): Initial 4 layer reservoirs with 100 neurons in each reservoir on L Z dataset; (d): Initial 8 layer reservoirs with 50 neurons in each reservoir on L Z dataset.
Applsci 13 02918 g009
Figure 10. Testing NRMSE comparison of unpruned DeepESN and DeepESN pruned by IPMA, ED-IPMA, PC-IPMA, SC-IPMA, KC-IPMA. (a): Initial 4 layer reservoirs with 100 neurons in each reservoir on M G dataset; (b): Initial 8 layer reservoirs with 50 neurons in each reservoir on M G dataset; (c): Initial 4 layer reservoirs with 100 neurons in each reservoir on L Z dataset; (d): Initial 8 layer reservoirs with 50 neurons in each reservoir on L Z dataset.
Figure 10. Testing NRMSE comparison of unpruned DeepESN and DeepESN pruned by IPMA, ED-IPMA, PC-IPMA, SC-IPMA, KC-IPMA. (a): Initial 4 layer reservoirs with 100 neurons in each reservoir on M G dataset; (b): Initial 8 layer reservoirs with 50 neurons in each reservoir on M G dataset; (c): Initial 4 layer reservoirs with 100 neurons in each reservoir on L Z dataset; (d): Initial 8 layer reservoirs with 50 neurons in each reservoir on L Z dataset.
Applsci 13 02918 g010
Table 1. Minimum testing NRMSE condition of each experiment group.
Table 1. Minimum testing NRMSE condition of each experiment group.
GroupUnpruned DeepESNPruned DeepESN
Method-ED-IPMAPC-IPMASC-IPMAKC-IPMAIPMA
N R M S E t e s t (Mean)0.036 1050.036 0460.036 2400.036 2060.036 2430.036 380
M G : 4 × 100 N R M S E t e s t (Std.)0.000 4830.000 3880.000 2210.000 2810.000 1440.000 291
Neuron number12012080120120120
N R M S E t e s t (Mean)0.035 5440.035 2380.035 3500.035 5160.035 4270.036 030
M G : 8 × 50 N R M S E t e s t (Std.)0.000 6990.000 3460.000 5160.000 5330.0007370.000 338
Neuron number160200240280280240
N R M S E t e s t (Mean)0.184 2850.148 8120.152 7750.164 4550.140 9300.155 667
L Z : 4 × 100 N R M S E t e s t (Std.)0.043 2010.026 0720.024 4170.022 9530.031 8620.027 816
Neuron number400200160280240200
N R M S E t e s t (Mean)0.251 5350.198 5810.189 2300.210 3760.193 6780.186 933
L Z : 8 × 50 N R M S E t e s t (Std.)0.044 4840.027 6310.024 9950.033 4040.029 1640.024 092
Neuron number320160120160160120
Red bold values indicate the best validation performance of all methods.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shen, Q.; Zhang, H.; Mao, Y. Improving Deep Echo State Network with Neuronal Similarity-Based Iterative Pruning Merging Algorithm. Appl. Sci. 2023, 13, 2918. https://doi.org/10.3390/app13052918

AMA Style

Shen Q, Zhang H, Mao Y. Improving Deep Echo State Network with Neuronal Similarity-Based Iterative Pruning Merging Algorithm. Applied Sciences. 2023; 13(5):2918. https://doi.org/10.3390/app13052918

Chicago/Turabian Style

Shen, Qingyu, Hanwen Zhang, and Yao Mao. 2023. "Improving Deep Echo State Network with Neuronal Similarity-Based Iterative Pruning Merging Algorithm" Applied Sciences 13, no. 5: 2918. https://doi.org/10.3390/app13052918

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop