Next Article in Journal
A High-Gain Multiphase Interleaved Differential Capacitor Clamped Boost Converter
Previous Article in Journal
An N/M-Ratio All-Digital Clock Generator with a Pseudo-NMOS Comparator-Based Programmable Divider
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Learning Method for Stochastic Configuration Networks Using Non-Inverse Solution

1
School of Information and Control Engineering, China University of Mining and Technology, Xuzhou 221116, China
2
Key Laboratory of Synthetical Automation for Process Industries, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(2), 262; https://doi.org/10.3390/electronics11020262
Submission received: 1 November 2021 / Revised: 19 December 2021 / Accepted: 22 December 2021 / Published: 14 January 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
Stochastic configuration networks (SCNs) face time-consuming issues when dealing with complex modeling tasks that usually require a mass of hidden nodes to build an enormous network. An important reason behind this issue is that SCNs always employ the Moore–Penrose generalized inverse method with high complexity to update the output weights in each increment. To tackle this problem, this paper proposes a lightweight SCNs, called L-SCNs. First, to avoid using the Moore–Penrose generalized inverse method, a positive definite equation is proposed to replace the over-determined equation, and the consistency of their solution is proved. Then, to reduce the complexity of calculating the output weight, a low complexity method based on Cholesky decomposition is proposed. The experimental results based on both the benchmark function approximation and real-world problems including regression and classification applications show that L-SCNs are sufficiently lightweight.

1. Introduction

Although the deep neural networks have proven to be a powerful learning tool, most networks suffer from time-consuming training due to the massive hyperparameters and complex structures. In many heterogeneous data analytics tasks, flattened networks can achieve promising performance. In the flattened networks, single-hidden layer feedforward neural networks (SLFNs) [1,2] have been widely applied because of their universal approximation capability and simple construction. However, gradient-descent-based learning algorithms are generally adopted for SLFNs training. Therefore, slow convergence and trap in a local minimum are often-encountered problems [3].
The randomized learning method offers a different learning method for flattened networks training. Many randomized flattened networks have been shown to approximate continuous functions on compact sets, and they also have the property of fast learning [4,5]. Stochastic configuration networks (SCNs) [6] provide a state-of-the-art randomized incremental learning method for SLFNs. In comparison with the traditional randomized incremental learning models, SCNs have some advantages: (1) SCNs randomly assign the input weights and biases of the hidden nodes in dynamically adjustable scopes according to the supervisory mechanism; (2) a more compact network structure. Therefore, SCNs have been extensively studied and become a hot topic of neural computing.
For large-scale data analytics, an ensemble learning method for quickly disassociating heterogeneous neurons was proposed is designed for SCNs by using a negative correlation learning strategy [7]. To improve learning efficiency, SCNs with block increments and variable increments are developed, which allow multiple hidden nodes to be added at each iteration [8,9]. Then, point and block increments are integrated into the parallel SCNs (PSCNs) [10]. To resolving the modeling tasks of uncertain data, robust SCN (RSCNs) is proposed by using maximum correlation entropy criterion (MCC) and kernel density estimation [10,11,12]. In order to further improve the expressiveness, SCNs with deep and stacked structures are proposed [13,14,15]. In [16], a two-dimensional SCNs (2DSCNs) is constructed for image data analytics. To address prediction interval estimation problems, the corresponding deep, ensemble, robust, and sparse versions of SCNs were developed [17,18,19]. In addition to the above theoretical studies, SCNs have successful applications in many fields, such as optical fiber pre-warning system [20], industrial process [21], concrete defect recognition [22], and so on.
However, the SCNs construction process can be extremely time-consuming when dealing with complex modeling tasks. The fundamental reason for this is that singular value decomposition (SVD) is needed to solve the output weights [23]. Concretely, the number of rows of the hidden layer output matrix is always much larger than the number of columns [24,25], which makes the hidden layer output matrix to become an over-determined matrix with no inverse [26]. To obtain the output weights, it is necessary to employ SVD to solve Moore–Penrose (M–P) generalized inverse in an over-determined matrix. Theoretically, the complexity of SVD is related with third power of the number of hidden nodes and the product of number of hidden nodes and input dimension. This makes the modeling process of SCNs extremely time-consuming when dealing with complex tasks that require a large network structure (a large number of hidden nodes) to enhance the expressive power of the model.
This paper proposes a lightweight non-inverse solution method for the output weights of SCNs (L-SCNs) by introducing normal equation theory [27] and Cholesky decomposition [28]. The main contributions of the paper are as follows:
  • To avoid adopting M–P generalized inverse with SVD, a positive definite equation for solving output weights is established based on normal equation theory to replace the over-determined equation;
  • The consistency of the solutions of the positive definite equation and the over-determined equation in calculating the output weights is proved;
  • A low complexity method for solving the positive definite equations based on Cholesky decomposition is proposed.
Experimental results on both the benchmark function approximation and real-world problems including regression and classification applications show that, compared with SCNs and IRVFLNs (an incremental variant of RVFLNs), the proposed L-SCNs have a superior performance in lightweight aspect.
The remaining parts of the paper are organized as follows. In Section 2, the basic principle of SCNs and some remarks are shown. The algorithm description of L-SCNs and full proof of related theories are presented in Section 3. In Section 4, the experimental setup is given and the performance of L-SCNs is fully discussed. Some conclusions are drawn in Section 5.

2. Brief Review of SCNs

As a kind of flattened network, SCNs model includes an input layer, hidden layer and output layer. Its hidden layer is constructed incrementally according to the supervisory mechanism. The specific SCNs network structure is shown in Figure 1.
The construction process of SCNs is briefly described as follows:
Given an input X = x 1 , x 2 , , x N , x i d and its corresponding output f = f 1 , f 2 , , f N , f i m . Suppose that we have already built a SCNs with L−1 hidden nodes, i.e.,:
f L 1 x = j = 1 L 1 β j g j w j T x + b j
where β j = β j , 1 , β j , 2 , , β j , m T is the output weights vector of the j-th hidden node, w j is the input weight vector of the j-th hidden node. b j is the threshold of the j-th hidden node. g j w j T x + b j is the hidden layer output vector of the j-th node, “ T ” represents a transpose of the matrix.
The current residual error of SCNs is calculated by Equation (2):
e L 1 = f f L 1
The acceptable tolerance error, denoted as ε. If eL−1 does not reach ε, continue to add new nodes to the SCNs by the supervision configuration mechanism:
ξ L = q = 1 m ξ L , q 0
and
ξ L , q = e L 1 , q T h L 2 h L T h L 1 r μ L e L 1 , q T 2 , q = 1 , 2 , , m
where 0 < r < 1 indicates regularization parameter, μ L is a nonnegative real number sequence with μ L 1 r and lim L μ L = 0 , h L = g L w L T x 1 + b L , , g L w L T x N + b L T .
The best hidden node parameters are determined by the maximum ξ L . Then, the output weights can be evaluated by Equation (5):
β = H L f
where β = β 1 , , β L . H L = h 1 , , h L is the current hidden layer output matrix, h p = g p , p = 1 , , L . H L is M–P generalized inverse matrix of H L .
The above process will be repeated until the residual error reach expected a tolerance ε or hidden nodes reaches the maximum.
Remark 1.
It can be seen that the hidden nodes of SCNs are built incrementally, and all output weights need to be recalculated after each hidden node is added. Therefore, the complexity of the modeling process depends on the evaluation method of the output weights.
Remark 2.
It can be seen from the above analysis, H L is an over-determined matrix. Therefore, M–P generalized inverse method is used to solve the over-determined equation H L β = f . However, the complexity of M–P generalized inverse method is related with third power of the number of hidden nodes and the product of number of hidden nodes and input dimension due to the use of singular value decomposition (SVD). Thus, the M–P method is very time-consuming, especially when dealing with complex modeling tasks that require a large number of hidden nodes. In addition, the M–P generalized inverse method can only obtain the approximate solution of the output weights, which is difficult to make the model optimal.

3. L-SCNs Method

From the above analysis, it can be seen that the M–P generalized inverse method involving SVD is the main reason for time-consuming nature of SCNs modeling. In order to solve this problem, a positive definite equation based on normal equation is proposed to replace the over-determined equation. Then, a low computational complexity method based on Cholesky decomposition is proposed to solve the positive definite equation and obtain the output weight, thereby reducing the modeling complexity of SCNs.

3.1. Positive Definite Equation

For the sake of brevity, this paper introduces H to replace HL. According to normal equation theory, H β = f can be denoted by
H T H β = H T f
Theorem 1 can guarantee the consistency of the solution of positive definite equation and over-determined equation, and a strict proof is given.
Theorem 1.
The necessary and sufficient condition for β* is the least square solution of H β = f : β* is the solution of H T H β = H T f .
Proof. 
Sufficiency: Suppose an N-dimensional vector β* such that H T H β = H T f , given any n-dimensional vector β, β β . Let y = β β , y 0 , so.
f H β 2 2 = f H β H y 2 2 = f H β H y , f H β H y = f H β , f H β 2 H y , f H β + H y , H y = f H β 2 2 + H y 2 2 f H β 2 2
Therefore, β is a solution H β = f .
Necessity: Let r = f H β , the i-th component of r can be written as
r i = f i k = 1 n h i k β k , i = 1 , 2 , , m
Let
J = J β 1 , β 2 , , β n = r 2 2 = i = 1 m b i k = 1 n h i k β k 2
From the necessary conditions of the extreme value of the multivariate function, it can be obtained
J β j = 2 i = 1 m b i k = 1 n h i k β k h i j = 0 , j = 1 , 2 , , n
that is
k = 1 n i = 1 m h i j h i k β k = i = 1 m h i j y i , j = 1 , 2 , , n
Equation (11) can be transformed into matrix form:
H T H β = H T f
Based on the above analysis, the solutions of the two equations are consistent in theory. □

3.2. SCNs with Cholesky Decomposition

In order to reduce the computational complexity of the model, this paper uses the Cholesky decomposed method that does not involve the inversion operation to solve Equation (12). However, using Cholesky decomposition has a premise that the decomposed matrix must be a positive definite symmetric matrix. In addition, since H is not always full rank in practical applications, H T H is not necessarily a positive definite matrix. In this paper, we introduce a moderator factor I / C to make H T H a full rank matrix. I is the identity matrix of the same type as H T H , and C is determined by cross verification. Thus, Equation (6) can be denoted by H T H + I / C β = H T f . Let H T H + I / C = A , H T f = b , we have
A β = b
The transpose of A can be evaluated by Equation (14)
A T = H T H + I / C T = H T H + I / C
therefore, A = A T . A is a symmetric matrix.
Given an arbitrary vector v 0 , then the quadratic form of A can be expressed as
v T A v = v T H T H + I / C v = I / C v T v + H v T H v > 0
Based on the results, it is easy to verify that A is a positive definite symmetric matrix.
The solving process of β* based on Cholesky decomposition is as follows:
First, A is decomposed by
A = S S T
Let S = s 11 0 0 s 21 s 22 0 s L 11 s L 12 s L 1 L 1 , A = a 11 a 12 a 1 L 1 a 21 a 22 a 2 L 1 a L 11 a L 12 a L 1 L 1 . Based on Equation (16), the element sij in S that is not 0 can be evaluated by
s i j = a i i n = 1 i 1 s i n 2 i = j a i j n = 1 i 1 s i n s j n / s j j i > j
where i , j = 1 , 2 , , L 1 .
Bring Equation (16) into Equation (13), and multiply both sides of the formula by S 1 , then it can get
S T β = K
where S K = b .
Therefore, Equation (18) can be denoted by S K = b . The element calculation method in K is evaluated by:
k i = b i / s i i i = 1 b i n = 1 i 1 s n i k n / s i i i > 1
To sum up, the output weights β i can be calculated by:
β i = k i / s i i i = L 1 k i n = 1 L 1 i s i + n , i β i + n / s i i i < L 1
The pseudo code of L-SCNs is described in Algorithm 1:
Algorithm 1 L-SCNs
Inputs: X = x 1 , x 2 , , x N , x i d
Outputs: f = f 1 , f 2 , , f N , f i m
Initialization parameters: T max as the maximum times of random configuration, L max as the maximum number of hidden nodes, ε as the error tolerance, γ = λ min : λ : λ max
1. Initialization: e 0 = f , set 0 < r < 1 and Ω , W = [   ] , L = 1
2. While L L m a x or e 0 > ε , Do
 (1). Hidden Node Parameters Configuration (3–20)
3. For λ γ , Do
4.  For k = 1 , 2 , , T max , Do
5.   Randomly assign hidden nodes ( w L , b L ) from λ , λ d and λ , λ , respectively,
6.  Calculate h L based on h L = 1 / 1 + exp w L x k T b L , set μ L = 1 r / L + 1 and calculate ξ L , q by Equation (3)
7.  If min ξ L , 1 , , ξ L , 2 , , ξ L , m 0
8.   Save w L , b L in W, and ξ L in Ω
9.   Else
10.    go to back to step 4
11.   End If
12.  End For (step4)
13.  If W is not empty
14.   Find w L , b L that maximize ξ L in Ω
15.  Set H L = h 1 , h 2 , , h L
16.   Break (step 21)
17.  Else
18.    Randomly take τ 0 , 1 r and let r = r + τ
19.  End If
20. End For (step 3)
 (2). Evaluate the Output Weights (21–28)
21. Obtain H L = h 1 , h 2 , , h L
22. Calculate A by Equation (14)
23. Calculate S by Equation (16)
24. Calculate β by Equations (18)–(20)
25. Calculate e L = f H L β
26.  Update e 0 = e L , L = L + 1
27. End While
28. Return β = β 1 , , β L , w = w 1 , , w L , b = b 1 , , b L

3.3. Computational Complexity Analysis

It can be seen from the above description that the difference between the two methods lies in the calculation of the output weights β . SCNs obtains the output weights by the product of M–P generalized inverse matrix and the output f , while L-SCNs evaluates the output weights by positive definite equation and Cholesky decomposition, since the M–P generalized inverse is calculated using the SVD method. Therefore, the computational complexity of the output weights of SCNs is about O L 3 + L M d . While L-SCNs only involves simple addition, subtraction, multiplication, and division operations when calculating output weights, so the computational complexity is about O L 3 / 3 + L M d + L 2 M . Where M is the number of samples in the training set of classification, and d is the number of categories (d = 1 in the regression problem). In summary, the method proposed in this paper has obvious lightweight advantages when dealing with complex tasks that require a large number of hidden nodes.

4. Experiments

In this section, the performance of L-SCNs is evaluated and compared with original SCNs and IRVFLNs on some benchmark data sets. The sigmoid function is used as activation function. All experiments on L-SCNs, SCNs and IRVFLNs are performed in the MATLAB 2019b environment running on a Windows personal computer with Intel(R) Xeon(R) E3-1225 v6 3.31GHz CPUs and RAM 32 GB.

4.1. Data Sets Description

Eight data sets have been used in experiments, including five real regression problems and three real classification problems, which were collected from KEEL and UCI HAR. (Knowledge Extraction based on Evolutionary Learning (KEEL) [29] and UCI HAR database [30]). These data sets specifically include winequality-white, California, delta_ail, Compactiv, Abalone, Iris, Human Activity Recognition (HAR) and wine. In addition, there is a highly nonlinear benchmark regression function data set [31,32], which is generated by Equation (21). The detailed information of all the data sets are shown in Table 1.
f x = 0.2 e 10 x 4 2 + 0.5 e 80 x 40 2 + 0.3 e 80 x 20 2
where input x 0 , 1 and output f x are normalized to [−1, 1].

4.2. Experimental Setup

In each trial, all samples were randomly divided into training and test data sets. All the results in the paper are average of 30 trials on the data set. The specifications of the experimental setup are shown in Table 2, in which ε is the expected error tolerance, T max is the maximum times of random configuration, L max is the maximum number of hidden nodes. γ is the assignment range of hidden layer node parameters. The moderator factor C was obtained by cross validation.

4.3. Performance Comparison

First of all, the convergence and function fitting performance of IRVFLNs, SCNs, and L-SCNs are evaluated using a highly nonlinear benchmark regression function dataset. The results shown in Table 3, includes training time, training error, testing error and the number of hidden nodes, and the best experimental results are highlighted. It can be seen from Table 3 that the modeling times of L-SCNs are 18.8% and 66.69% lower than that of SCNs and IRVFLNs, respectively. The training error and testing error of L-SCNs have obvious advantages, especially compared with IRVFLNs. In addition, compared with IRVFLNs, SCNs and L-SCNs save 36.8% and 43.83% of hidden nodes, respectively. This is mainly because the hidden node parameter selection function of the supervision mechanism improves the compactness of the model while ensuring the high performance of the model. Since SCNs can only obtain approximate solutions when using M–P generalized inverse to calculate output weights, while L-SCNs output weight evaluation method can get real solutions. Therefore, L-SCNs is superior to SCNs in compactness and model performance.
In addition, in order to analyze the convergence and fitting ability of IRVFLNs, SCNs and L-SCNs, this paper draws a convergence curve and a fitting curve, as shown in Figure 2. It can be seen from the convergence curve that IRVFLNs used up 100 preset hidden nodes, but still did not meet the expected error tolerance. In particular, it is difficult to improve the convergence of IRVFLNs by adding more nodes after the number of hidden nodes reaches 51. The convergence of SCNs and L-SCNs meets the expected error tolerance, and L-SCNs converges faster. It only uses 19 nodes to reduce the residual to 0.02, and only used 56.17 nodes to meet the expectations. Compared with SCNs, L-SCNs save 11.12% of nodes. Therefore, it shows that L-SCNs modeling is faster, and the structure of the built model is more compact. The fitting curve shows that among IRVFLNs, SCNs and L-SCNs, the data fitting ability of the model built by IRVFLNs is the worst, while the models built by SCN and L-SCN have similar fitting capabilities.
IRVFLNs, SCNs and L-SCNs. The experimental results are presented Table 4 and Table 5, respectively. Table 4 and Table 5 show the experimental results of the regression problem and the classification problem, respectively. For each model, the best experimental results are highlighted in Table 4 and Table 5. Table 4 gives the number of hidden nodes, the training time, the training error and the testing error. It can be seen from Table 4 that for Abalone data set, the training error and test error of IRVFLNs are the worst, and 100 hidden nodes are used up, which is also the main reason for the longest modeling time. L-SCNs and SCNs achieve similar training error and testing error, but L-SCNs saves 73.18% and 25.77% of the number of nodes and modeling time, respectively. On the Compactiv data set, IRVFLNs still used up all hidden nodes, and achieved the worst training error and testing error. The experimental results of L-SCNs and SCNs are also consistent with the results on the Abalone data set. By comparing the experimental results of winequality-white, california and delta_ail, it can be seen that: (1) When consuming the same hidden layer node, IRVFLNs modeling is the fastest, but the model performance is the worst; (2) The number of hidden nodes required for L-SCNs modeling is far less than that of the other two algorithms; (3) When the number of hidden layer nodes is small, L-SCNs has no obvious advantage in lightness. In summary, L-SCNs are superior to IRVFLN and SCNs in terms of model compactness and modeling time when a large number of hidden layer nodes are needed.
Table 5 shows the numbers of hidden node, training times, training errors and testing errors of IRVFLNs, SCNs and L-SCNs on the three real classification data sets. As can be seen from Table 5, for the Iris data set, the numbers of hidden nodes required by SCNs and L-SCNs is much less than 107.2 of IRVFLNs. Therefore, SCNS and L-SCNs have lower modeling times. The main reason behind this result is that the node parameter selection function of the supervisory mechanism makes the node parameters quality better, so the model can reach the expected value faster and perform better. Compared with SCNs, L-SCNs saves 3.58% and 5.35% in the number of hidden nodes and training time, respectively. At the same time, L-SCNs achieves the best test error. In particular, the training errors of IRVFLNs, SCNs and L-SCNs on the Iris data set is the same. For the HAR data set, compared to the other two algorithms, L-SCNs saves 74.85% and 4.91% in the number of hidden nodes, while saving 73.01% and 40.04% in training time, respectively. In addition, the number of hidden nodes of IRVFLNs reached the maximum value of 1000, but the model performance was the worst. For the wine data set, L-SCNs and SCNs still have obvious advantages in the number of hidden nodes, training time and test error. Compared with the other two algorithms, L-SCNs constructs the best performance model with the least 90 hidden nodes and the minimum 0.1722 s training time. In summary, L-SCNs have obvious merits in training efficiency and model compactness for classification tasks. Therefore, L-SCNs is a lightweight algorithm. Through the analysis of Table 1 and Table 5, it can be seen that HAR and wine data sets have higher sample numbers and features than Iris data set, especially HAR data set. The experimental results also show that L-SCNs have more obvious in lightweight on HAR and wine data sets. Therefore, L-SCNs are suitable for dealing with large data problems.
In order to further verify the advantages of L-SCNs in terms of lightweight. In this paper, while maintaining the same number of hidden nodes, the change process of modeling time of SCNs and L-SCNs with the increase of the number of hidden nodes is drawn when the experiment is performed on the HAR data set, as shown in Figure 3. It can be seen that before the hidden nodes reach 100, the modeling time of SCNs and L-SCNs is the same. However, after 100 hidden nodes, with the increase of hidden nodes, the advantage of L-SCNs becomes more and more obvious in term of lightweight. When 500 nodes are reached, the gap between SCNs and L-SCNs widened to 36.66%. It also shows that when dealing with modeling tasks that require a large number of hidden nodes, the L-SCNs proposed in this paper can effectively reduce the modeling complexity and improve the lightweight of modeling.
In addition, we have compared the Cholesky decomposition approach with other methods, including QR decomposition, LDL decomposition and SVD decomposition; the detailed results of all these approaches are shown in Table 6. It can be found from Table 6 that Cholesky decomposition is slightly better than QR decomposition and LDL decomposition. However, as the number of nodes increases, compared with SVD decomposition, Cholesky decomposition has more obvious advantages in terms of lightness. The main reason for this result is that the computational complexity of QR and LDL decomposition is similar to that of Cholesky decomposition. The computational complexity of SVD far exceeds these three methods. This clearly demonstrates the lightness of Cholesky decomposition.

5. Conclusions

This work is motivated by the time-consuming calculation of output weights in each addition of hidden nodes. Lightweight stochastic configuration networks (L-SCNs) are developed by employing a non-inverse calculation method for problem solving. In L-SCNs, a positive definite equation is firstly proposed based on normal equation theory to take the place of the over-determined equation to avoid the use of M–P generalized inverse. Secondly, the Cholesky decomposition method with low computational complexity is used to calculate the positive definite equation and obtain the output weight. The proposed L-SCNs have been evaluated on several benchmark data sets, and the experimental results show that L-SCNs not only solve the high complexity problem of calculating output weights, but also improve the compactness of the model structure. In addition, the comparison with IRVFLNs and SCNs shows that L-SCNs have obvious advantages in lightweight. Therefore, L-SCNs are particularly suitable for complex modeling tasks that usually require a mass of hidden nodes to build an enormous network.

Author Contributions

Conceptualization, formal analysis, methodology, and writing-original draft, J.N.; data curation, Z.J. and C.N. writing-review and editing, W.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 61973306, in part by the Nature Science Foundation of Jiangsu Province under Grant BK20200086, in part by the Open Project Foundation of State Key Laboratory of Synthetical Automation for Process Industries under Grant 2020-KF-21-10.

Data Availability Statement

The data sets used in this paper are from UCI data sets (http://archive.ics.uci.edu/ml/index.php (accessed on 21 December 2021)), KEEL (https://sci2s.ugr.es/keel/category.php?cat=clas (accessed on 21 December 2021)) data sets, etc.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Han, F.; Jiang, I.; Ling, Q.H.; Su, B.H. A survey on metaheuristic optimization for random single-hidden layer feedforward neural network. Neurocomputing 2019, 335, 261–273. [Google Scholar] [CrossRef]
  2. Tamura, S.; Tateishi, M. Capabilities of a four-layered feedforward neural network: Four layers versus three. IEEE Trans. Neural Netw. 1997, 8, 251–255. [Google Scholar] [CrossRef] [PubMed]
  3. Wu, X.; Rozycki, P.; Wilamowski, B.M. A Hybrid Constructive Algorithm for Single-Layer Feedforward Networks Learning. IEEE Trans. Neural Netw. Learn. Syst. 2017, 26, 1659–1668. [Google Scholar] [CrossRef] [PubMed]
  4. Igelnik, B.; Pao, Y.H. Stochastic choice of basis functions in adaptive function approximation and the functional-link net. IEEE Trans. Neural Netw. 1995, 6, 1320–1329. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Pao, Y.H.; Takefuji, Y. Functional-link net computing: Theory, system architecture, and functionalities. Computer 1992, 25, 76–79. [Google Scholar] [CrossRef]
  6. Wang, D.H.; Li, M. Stochastic Configuration Networks: Fundamentals and Algorithms. IEEE Trans. Cybern. 2017, 47, 3466–3479. [Google Scholar] [CrossRef] [Green Version]
  7. Wang, D.H.; Cui, C.H. Stochastic Configuration Networks Ensemble for Large-Scale Data Analytics. Inf. Sci. 2017, 417, 55–71. [Google Scholar] [CrossRef] [Green Version]
  8. Dai, W.; Li, D.P.; Zhou, P.; Chai, T.Y. Stochastic configuration networks with block increments for data modeling in process industries. Inf. Sci. 2019, 484, 367–386. [Google Scholar] [CrossRef]
  9. Tian, Q.; Yuan, S.J.; Qu, H.Q. Intrusion signal classification using stochastic configuration network with variable increments of hidden nodes. Opt. Eng. 2019, 58, 026105.1–026105.8. [Google Scholar] [CrossRef]
  10. Dai, W.; Zhou, X.Y.; Li, D.P.; Zhu, S.; Wang, X.S. Hybrid Parallel Stochastic Configuration Networks for Industrial Data Analytics; IEEE Transactions on Industrial Informatics: Piscataway, NJ, USA, 2021. [Google Scholar] [CrossRef]
  11. Wang, D.H.; Li, M. Robust Stochastic Configuration Networks with Kernel Density Estimation for Uncertain Data Regression. Inf. Sci. 2017, 412, 210–222. [Google Scholar] [CrossRef]
  12. Li, M.; Huang, C.Q.; Wang, D.H. Robust stochastic configuration networks with maximum correntropy criterion for uncertain data regression. Inf. Sci. 2018, 473, 73–86. [Google Scholar] [CrossRef]
  13. Wang, D.H.; Li, M. Deep Stochastic Configuration Networks: Universal Approximation and Learning Representation. In Proceedings of the IEEE International Joint Conference on Neural Networks, Anchorage, AK, USA, 14–19 May 2017. [Google Scholar]
  14. Pratama, M.; Wang, D.H. Deep Stacked Stochastic Configuration Networks for Non-Stationary Data Streams. Inf. Sci. 2018, 495, 150–174. [Google Scholar] [CrossRef] [Green Version]
  15. Lu, J.; Ding, J.L. Construction of prediction intervals for carbon residual of crude oil based on deep stochastic configuration networks. Inf. Sci. 2019, 486, 119–132. [Google Scholar] [CrossRef]
  16. Li, M.; Wang, D.H. 2-D Stochastic Configuration Networks for Image Data Analytics. IEEE Trans. Cybern. 2021, 51, 359–372. [Google Scholar] [CrossRef]
  17. Lu, J.; Ding, J.L.; Dai, X.W.; Chai, T.Y. Ensemble Stochastic Configuration Networks for Estimating Prediction Intervals: A Simultaneous Robust Training Algorithm and Its Application. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 5426–5440. [Google Scholar] [CrossRef] [PubMed]
  18. Lu, J.; Ding, J.L.; Liu, C.X.; Chai, T.Y. Hierarchical-Bayesianbased sparse stochastic configuration networks for construction of prediction intervals. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–2. [Google Scholar] [CrossRef]
  19. Lu, J.; Ding, J.L. Mixed-distribution-based robust stochastic configuration networks for prediction interval construction. IEEE Trans. Ind. Inform. 2020, 16, 5099–5109. [Google Scholar] [CrossRef]
  20. Sheng, Z.Y.; Zeng, Z.Q.; Qu, H.Q.; Zhang, Y. Optical fiber intrusion signal recognition method based on TSVD-SCN. Opt. Fiber Technol. 2019, 48, 270–277. [Google Scholar] [CrossRef]
  21. Xie, J.; Zhou, P. Robust Stochastic Configuration Network Multi-Output Modeling of Molten Iron Quality in Blast Furnace Ironmaking. Neurocomputing 2020, 387, 139–149. [Google Scholar] [CrossRef]
  22. Zhao, J.H.; Hu, T.Y.; Zheng, R.F.; Ba, P.H. Defect Recognition in Concrete Ultrasonic Detection Based on Wavelet Packet Transform and Stochastic Configuration Networks. IEEE Access 2021, 99, 9284–9295. [Google Scholar] [CrossRef]
  23. Salmerón, M.; Ortega, J.; Puntonet, C.G.; Prieto, A. Improved RAN sequential prediction using orthogonal techniques. Neurocomputing 2001, 41, 153–172. [Google Scholar] [CrossRef]
  24. Qu, H.Q.; Feng, T.L.; Zhang, Y. Ensemble Learning with Stochastic Configuration Network for Noisy Optical Fiber Vibration Signal Recognition. Sensors 2019, 19, 3293. [Google Scholar] [CrossRef] [Green Version]
  25. Liu, J.; Hao, R.; Zhang, T.; Wang, X.Z. Vibration fault diagnosis based on stochastic configuration neural networks. Neurocomputing 2021, 434, 98–125. [Google Scholar] [CrossRef]
  26. Krein, S.G. Overdetermined Equations; Birkhäuser: Basel, Switzerland, 1982. [Google Scholar]
  27. Loboda, A.V. Determination of a Homogeneous Strictly Pseudoconvex Surface from the Coefficients of Its Normal Equation. Math. Notes 2003, 73, 419–423. [Google Scholar] [CrossRef]
  28. Roverato, A. Cholesky decomposition of a hyper inverse Wishart matrix. Biometrika 2000, 87, 99–112. [Google Scholar] [CrossRef]
  29. Anguita, S.; Ghio, A.; Oneto, L.; Parra, X. Energy efficient smartphone-based activity recognition using fixed-point arithmetic. J. Univers. Comput. 2013, 19, 1295–1314. [Google Scholar]
  30. Fdez, J.A.; Fernandez, A.; Luengo, J.; Derrac, J.; Garacia, S.; Herrera, F. KEEL Data-Mining Software Tool: Data Set Repository, Integration of Algorithms and Experimental Analysis Framework. J. Mult.-Valued Log. Soft Comput. 2011, 17, 255–287. [Google Scholar]
  31. Tyukin, I.Y.; Prokhorov, D.V. Feasibility of random basis function approximators for modeling and control. In Proceedings of the IEEE Control Applications, (CCA) & Intelligent Control, St. Petersburg, Russia, 8–10 July 2009; pp. 1391–1396. [Google Scholar]
  32. Li, M.; Wang, D.H. Insights into randomized algorithms for neural networks: Practical issues and common pitfalls. Inf. Sci. 2017, 382, 170–178. [Google Scholar] [CrossRef]
Figure 1. Network structure of SCNs. 1, 2, j, j + 1 represent the node 1, 2, j, j + 1 respectively. d represents the dimension of the input data set and m represents the dimension of the output data.
Figure 1. Network structure of SCNs. 1, 2, j, j + 1 represent the node 1, 2, j, j + 1 respectively. d represents the dimension of the input data set and m represents the dimension of the output data.
Electronics 11 00262 g001
Figure 2. Training results of three algorithms. (a) Convergence Curve; (b) Fitting Curve.
Figure 2. Training results of three algorithms. (a) Convergence Curve; (b) Fitting Curve.
Electronics 11 00262 g002
Figure 3. Modeling time of SCNs and L-SCNs.
Figure 3. Modeling time of SCNs and L-SCNs.
Electronics 11 00262 g003
Table 1. Specifications of data sets.
Table 1. Specifications of data sets.
Data SetsNo. of SampleAttributesClasses
Training DataTest Data
Regressionnonlinear function8002001-
Abalone200021777-
Compactiv6144204821-
winequality-white3428147012-
delta_ail499021395-
california14,44861928-
ClassificationIris1203043
HAR735229475616
wine14236133
Table 2. Specifications of the experimental setup.
Table 2. Specifications of the experimental setup.
Data SetsExpected Error Algorithms   Parameters   L max , γ , T max
IRVFLNsSCNsL-SCNs
nonlinear function ε = 0.002 100, {1}, 1100, {150:10:200}, 20100, {150:10:200}, 20
Abalone ε = 0.16 100, {1}, 1100, {150:10:200}, 20100, {150:10:200}, 20
Compactiv ε = 0.15 200, {1}, 1200, {10:1:20}, 20200, {10:1:20}, 20
winequality-white ε = 0.05 100, {1}, 1100, {10:1:20}, 10100, {10:1:20}, 10
delta_ail ε = 0.21 100, {0.5}, 150, {0.5:0.1:10}, 1050, {0.5:0.1:10}, 10
california ε = 0.11 50, {1}, 150, {1:1:10}, 1050, {1:1:10}, 10
Iris ε = 0.05 200, {10}, 1100, {10:0.5:20}, 20100, {10:0.5:20}, 20
HAR ε = 0.05 500, {50}, 1500, {1:1:10}, 20500, {1:1:10}, 20
wine ε = 0.05 200, {0.5}, 1100, {0.5:0.5:10}, 20100, {0.5:0.5:10}, 20
Table 3. Performance comparison of highly nonlinear function.
Table 3. Performance comparison of highly nonlinear function.
ModelsLt (s)Training ErrorTesting Error
IRVFLNs1000.36570.07200.0714
SCNs63.200.15000.00160.0016
L-SCNs56.170.12180.00150.0014
Table 4. Performance comparison for regression data sets.
Table 4. Performance comparison for regression data sets.
Data SetsModelsLt (s)Training ErrorTesting Error
AbaloneIRVFLNs1000.27110.18950.1977
SCNs32.330.18740.15990.1641
L-SCNs8.670.13910.15900.1601
CompactivIRVFLNs2000.63850.17700.1862
SCNs270.22160.14650.1541
L-SCNs16.670.18560.14180.1501
winequality-whiteIRVFLNs1000.56040.23250.2472
SCNs1000.940.22760.2536
L-SCNs1000.890.22640.2487
californiaIRVFLNs23.570.120.10880.1101
SCNs16.330.140.11780.1171
L-SCNs10.330.140.11440.1155
delta_ailIRVFLNs27.670.090.20980.2163
SCNs11.330.100.20950.2113
L-SCNs9.330.120.20290.2107
Table 5. Performance comparison for classification data sets.
Table 5. Performance comparison for classification data sets.
Data SetsModelsLt (s)Training ErrorTesting Error
IrisIRVFLNs107.20.194700.1667
SCNs74.330.142400.0667
L-SCNs71.670.123700.0556
HARIRVFLNs100059.360.03500.0763
SCNs264.526.720.03220.0747
L-SCNs251.516.020.03220.0658
wineIRVFLNs117.80.387100.1611
SCNs97.330.261500.0695
L-SCNs90.000.172200.0667
Table 6. Comparison of different decomposition methods.
Table 6. Comparison of different decomposition methods.
Methods100200300400500
SVD6.33 s16.61 s35.47 s64.79 s107.78 s
LDL5.78 s14.00 s30.96 s49.54 s76.22 s
QR5.59 s13.69 s26.17 s44.03 s69.12 s
Cholesky5.53 s13.48 s25.83 s43.67 s68.27 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nan, J.; Jian, Z.; Ning, C.; Dai, W. A Lightweight Learning Method for Stochastic Configuration Networks Using Non-Inverse Solution. Electronics 2022, 11, 262. https://doi.org/10.3390/electronics11020262

AMA Style

Nan J, Jian Z, Ning C, Dai W. A Lightweight Learning Method for Stochastic Configuration Networks Using Non-Inverse Solution. Electronics. 2022; 11(2):262. https://doi.org/10.3390/electronics11020262

Chicago/Turabian Style

Nan, Jing, Zhonghua Jian, Chuanfeng Ning, and Wei Dai. 2022. "A Lightweight Learning Method for Stochastic Configuration Networks Using Non-Inverse Solution" Electronics 11, no. 2: 262. https://doi.org/10.3390/electronics11020262

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop