Next Article in Journal
Distributed Watchdogs Based on Blockchain for Securing Industrial Internet of Things
Next Article in Special Issue
Recent Advances in Touch Sensors for Flexible Wearable Devices
Previous Article in Journal
Robustly Adaptive EKF PDR/UWB Integrated Navigation Based on Additional Heading Constraint
Previous Article in Special Issue
A Multi-DoF Prosthetic Hand Finger Joint Controller for Wearable sEMG Sensors by Nonlinear Autoregressive Exogenous Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Channel Fusion Classification Method Based on Time-Series Data

1
School of Artificial Intelligent, Beijing Technology and Business University, Beijing 100048, China
2
China Light Industry Key Laboratory of Industrial Internet and Big Data, Beijing Technology and Business University, Beijing 100048, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(13), 4391; https://doi.org/10.3390/s21134391
Submission received: 14 April 2021 / Revised: 4 June 2021 / Accepted: 15 June 2021 / Published: 26 June 2021
(This article belongs to the Special Issue Wearable Sensor for Activity Analysis and Context Recognition)

Abstract

:
Time-series data generally exists in many application fields, and the classification of time-series data is one of the important research directions in time-series data mining. In this paper, univariate time-series data are taken as the research object, deep learning and broad learning systems (BLSs) are the basic methods used to explore the classification of multi-modal time-series data features. Long short-term memory (LSTM), gated recurrent unit, and bidirectional LSTM networks are used to learn and test the original time-series data, and a Gramian angular field and recurrence plot are used to encode time-series data to images, and a BLS is employed for image learning and testing. Finally, to obtain the final classification results, Dempster–Shafer evidence theory (D–S evidence theory) is considered to fuse the probability outputs of the two categories. Through the testing of public datasets, the method proposed in this paper obtains competitive results, compensating for the deficiencies of using only time-series data or images for different types of datasets.

1. Introduction

The development of sensor technology has increased storage capacity and equipment types and record a significant amount of time-series data. It is very important to perform time-series data analysis in, for instance, accurate classification processing, which is widely used to solve different practical problems, such as mobile object tracking [1], machine fault detection [2], and medical diagnosis [3].
Based on the investigation reported herein, it is found that there are two main time-series classification methods. The first mainly relies on the time series itself, using traditional machine learning or deep learning (DL) for classification. The second kind benefits from the development of image classification networks and encodes time series into images before classification. In this paper, both methods are considered to achieve the use of two modal features. Specifically, long short-term memory (LSTM), the gated recurrent unit (GRU), and bidirectional LSTM (BiLSTM) are selected as the feature extraction method for the original time series due to their ability in automatic feature extraction. Broad learning systems (BLSs) are selected for time-series images, which are simple and satisfy a BLS’s characteristic. In brief, in this paper, a multi-channel fusion classification model is presented to improve the classification effect for different types of series data.
The rest of this article is organized as follows. In Section 2, related work is introduced. In Section 3, the proposed model block diagram and detailed structure are presented. Section 4 presents the experimental data, experimental details, and analysis results. Finally, conclusions are drawn in Section 5.

2. Related Work

The processing of classification problems mainly depends on whether the data are similar or not. Time-series classification problems are also analyzed based on this concept. The method of extracting features can be divided into manual and automatic feature extraction for classification.

2.1. Methods Based on Manual Feature Extraction

Manual feature extraction is usually used in conjunction with traditional machine learning methods. Measures based on distance are generally adopted, such as Euclidean distance (ED) and dynamic time warping (DTW), and work with k-nearest-neighbor (KNN) classifiers [4]. Huang et al. [5] proposed a KNN algorithm based on class contribution and feature weighting that uses weighted ED to obtain k nearest neighbors. By using class contributions combining the number of k nearest neighbors and their average distance, the final predicted label of samples is obtained. This method achieves a high classification accuracy in tests of public datasets. However, when the data are deformed, such as by scaling, DTW works better than ED. The core idea of DTW is to automatically distort the time series—that is, to perform local scaling on the time axis—so that the shape of the two sequences is as consistent as possible to obtain the maximum possible similarity. Hu et al. [6] selected sample motion data and normalized it to create a template, and then they used the DTW method to compare the processed data with the template to achieve a higher accuracy of activity classification. Furthermore, DTW and ED can work together in the model. Do et al. [7] and Kurt et al. [8] both used the DTW method to align data and then calculated the ED value as one of the metrics for classification. In addition, the hidden Markov model (HMM) [9] and support vector machine (SVM) [10] are also effective machine learning methods. Wang et al. [11] used the Gaussian mixture model to fuse the extracted features and then used the HMM to estimate the output to classify electroencephalogram (EEG) signals. Alickovic et al. [12] proposed a classifier named RotSVM for sleep stage classification, in which the features after noise reduction and discrete wavelet transform are used as input. They built a model that can be effectively used in medical and home care applications.

2.2. Methods Based on Automatic Feature Extraction

Although machine learning methods show more superior performance in time-series data classification, many studies have shown that manual feature extraction is not easy with the growth of types and numbers of time series, and traditional machine learning is more suitable for sample learning with lower dimensions. As the superior performance of DL emerges, its application in the time-series analysis is gradually being explored for its ability of automatic feature extraction.
Recurrent neural networks (RNNs) [13,14] are the most commonly used method. Modelling of time-series data by an RNN considers the time correlation of data, which is reflected in the connection of nodes between hidden layers; that is, the input of the hidden layer includes not only the output of the input layer but also the output of the hidden layer at the previous time. In theory, an RNN can process sequence data of any length, but in practice, it is found that it cannot solve the long-term dependence problem. To maintain the memory and dependence on the data, RNN’s variants, LSTM and the GRU were proposed in turn. Dutta [13], compared the simple RNN, LSTM, and GRU with EEG signal data. As the number of layers increases, although it takes longer, the accuracy of the latter two is significantly higher than that of the former. Compared with LSTM, the training time of the GRU is shorter, but the accuracy of the two is comparable. RNN extension methods also include bidirectional LSTM and bidirectional GRU. The bidirectional structure allows the network to consider the context information of time series, and it can show very good results in some tasks, e.g., natural language processing.
In addition to RNN series methods, convolutional neural networks (CNNs) are also used for time-series classification. For example, Kong et al. [15] proposed a fine-grained visual recognition model called MCF-Net to classify different crop species in practical farmland scenes. With multi-stream hybrid architecture utilizing massive fine-granulometric information, MCF-Net obtains preferable representation ability for distinguishing interclass discrepancy and tolerating intra-class variances.
As far as the network architecture is concerned, the characteristics of DL networks are the vertical expansion of the network layer, which imposes a greater demand for computing resources, which, in turn, places higher requirements on hardware. Therefore, in recent years, networks aimed at improving training speed have gradually attracted researcher attention. Among them, BLSs provide an alternative method for DL networks, which also can extract features automatically. Based on a random vector functional link NN (RVFLNN) and incremental learning [16], Chen proposed the BLS [17]. As an efficient incremental learning system without a deep architecture, the wide network can classify images with low background complexity. Based on this finding, Yang et al. applied a BLS to the classification of time-series data and obtained a highly precise classification result [18].

2.3. Methods Based on Time-Series Encoding

The aforementioned methods are all from the perspective of data series, which need the memory capacity of the network or the similarity between data to be found through other methods to achieve time-series classification. With the development of DL in image classification, several researchers have discovered ways to encode data from the perspective of images and implement classification. Gramian angular field (GAF) and Markov transition field (MTF) methods proposed by Wang et al. [19] and the recurrence plot (RP) method proposed by Hatami et al. [20] all encode time-series data into images. The advantage is that the time relationship between data points can be directly displayed through images, and then the relationship could use the image classification networks for time-series classification. Inspired by this, Saeed et al. [21] used the GAF method and combined the Inception V3 model to achieve high-precision classification of time series.

3. Methods

The model framework of time-series data combined with multi-modal features presented in this paper mainly includes three parts: time-series data encoding and its feature extraction, original time-series data feature extraction, and decision-level fusion. The specific structural diagram is shown in Figure 1 and described in detail below.

3.1. Feature Extraction and Classification of Time-Series Images

In this subsection, the time series is first encoded to images by using RP and GAF, and then the BLS will be used to extract image features. A SoftMax layer is added to obtain the probability result for decision-level fusion.

3.1.1. RP Encoding Method

Inspired by the RP [22], Hatami et al. [20] used two-dimensional phase-space trajectories to visualize time series. RP can analyze the periodicity, chaos, and non-stationarity of time series, reveal the internal structure and give a priori knowledge about similarity, information, and predictability. It is especially suitable for short time-series data. The encoding process is the following.
First, given a time series X { x 1 x 2 , , x n } the time-delay embedding method is used to reconstruct the two-dimensional phase space. The state of the phase space with a time delay of 1 is expressed as:
s 1 : ( x 1 , x 2 ) , s 2 : ( x 1 , x 2 ) , , s n 1 : ( x n 1 , x n )
Then, the RP can be expressed as:
R i j = θ ( ε s i s j )
where θ ( ) denotes the Heaviside function, ε is the threshold, and is the norm; an infinite norm is usually used. In the actual encoding, to retain more image details through color transformation, θ ( ) is not used. In addition, if using an infinite norm for calculation, the image will appear symmetrical, which may make it difficult to distinguish some categories; thus, in this paper, the original difference value using the largest absolute difference value is brought into Equation (2) after comparison using the infinite norm value. Therefore, the RP is expressed as:
R i j = { ε ( x i 1 x j 1 ) , i f | x i 1 x j 1 | | x i 2 x j 2 | ε ( x i 2 x j 2 ) , i f | x i 1 x j 1 | < | x i 2 x j 2 |
where x i k   and   x j k ( k = 1 , 2 ) represent the k th value of vectors s i   and   s j , respectively. The visualization result is shown on the left-hand side of Figure 2.

3.1.2. GAF Encoding Method

The GAF method transfers the normalized series data to a polar coordinate system and then generates the Gramian angular summation field (GASF) or Gramian angular difference field (GADF) matrix by calculating the cosine and sine of the corresponding angle of each pair of elements and then displays the series data in the form of images. The specific conversion process is the following.
Given a time series X = { x 1 , x 2 , , x n } , since the type of data could be various, it is necessary to normalize the data to intervals [-1, 1] for the same dimension and reduce the amount of calculation,
x ˜ i = ( x i max ( X ) ) + ( x i min ( X ) ) max ( X ) min ( X )
For each piece of normalized data, the inverse cosine function is used to map to the polar coordinate system and process the time stamp as a radius; the formula is
{ ϕ i = a r c cos ( x ˜ i ) , 1 x ˜ i 1 , x ˜ i X ˜ r i = t i N , t i Ν
where t i is the time stamp, and N is the span of the constant polar coordinate system. In practical applications, its value is equal to the sequence length.
Then, GASF can be defined as:
G G A S F = [ cos ( ϕ 1 + ϕ 1 ) cos ( ϕ 1 + ϕ n ) cos ( ϕ 2 + ϕ 1 ) cos ( ϕ 2 + ϕ n ) cos ( ϕ n + ϕ 1 ) cos ( ϕ n + ϕ n ) ]
and GADF as:
G G A D F = [ sin ( ϕ 1 ϕ 1 ) sin ( ϕ 1 ϕ n ) sin ( ϕ 2 ϕ 1 ) sin ( ϕ 2 ϕ n ) sin ( ϕ n ϕ 1 ) sin ( ϕ n ϕ n ) ]
The above two matrices are used to obtain the images of sequence X, as shown on the right-hand side of Figure 2. Through polar coordinate conversion and trigonometric function mapping, the time correlation between different data points is directly displayed by the color of the image.

3.1.3. BLS

BLS has a variety of structural forms, and the classical structure shown in Figure 3 is used in this paper. It has two kinds of nodes: feature mapping nodes and enhancement nodes. The former performs nonlinear activation on the input data, while the latter, similar to the kernels in convolutional layers, is assumed to be used to fully exploit features in the data and improve the learning ability of the network. The design is the following.
First, the input data are subjected to feature mapping to form a feature node. Second, the feature nodes are enhanced to enhancement nodes by randomly generated weights. The optimal weight selection between the output layer and the feature and enhancement nodes can be obtained by ridge regression and pseudo-inverse algorithms. The specific process is the following.
Assuming that the input data is X with N samples, and each sample has M dimensions. Y is the output matrix that belongs to R N × C . The i th feature mapping groups are represented by
Z i = ϕ i ( X W e i + β e i )
where ϕ i is the nonlinear activation function, and W e i is fine-tuned using Equation (9) with iteration steps:
{ w k + 1 = ( Z T Z + ρ I ) 1 ( Z T x + ρ ( o k u k ) ) o k + 1 = S λ ρ ( w k + 1 + u k ) u k + 1 = u k + ( w k + 1 o k + 1 )
where ρ > 0 , and S is the soft threshold operator defined as follows,
S k ( a ) = { a k , a > k 0 , | a | k a + k , a < k
All generated feature nodes are represented by Z n [ Z 1 , , Z n ] , and then the m th group of the enhancement nodes is represented as:
H m ξ ( Z n W h m + β h m )
Therefore, the BLS model can be expressed as:
Y = [ Z 1 , , Z n | ξ ( Z n W h 1 + β h 1 ) , , ξ ( Z n W h m + β h m ) ] W m = [ Z 1 , , Z n | H 1 , , H m ] W m = [ Z n | H m ] W m
where W m = [ Z n | H m ] + Y ; W m is the connection weight of the broad structure, calculated by the ridge regression algorithm using the following formula to obtain the best value,
[ Z n | H m ] + = lim λ 0 ( λ I + [ Z n | H m ] [ Z n | H m ] T ) 1 [ Z n | H m ] T

3.2. Feature Extraction and Classification of Original Time-Series Data

In the preceding subsection, the images encoded from time-series data are used in classification, while the original time series is also considered in order to prevent the information learned from being insufficient.
Time series are limited or infinite data streams that depend on each other between data points, and an RNN is usually used to process such data. In this paper, LSTM, GRU, and BiLSTM are selected as the feature extraction methods of original series data in the way of parallelization, and the SoftMax layer is also added for the later operation of decision-level fusion. The structures of these three methods are introduced in the following subsections.

3.2.1. LSTM

As a special RNN, LSTM is mainly used to solve the problem of gradient disappearance and gradient explosion during long sequence training. In other words, LSTM can perform better in longer sequences than an ordinary RNN. The main reason is that LSTM adds a structure called a gate for selective control of the passage of information. Specifically, it includes three gates, called the forget, input, and output gates. The internal structure of LSTM is shown in Figure 4.
The forget gate is used to determine the retention of the information contained in the previous moment’s state. The input gate selects the new state information that must be added so as to obtain the state of the current moment. The output decides the final unit output at the current time. The Equations of the entire procedure are
f t = σ ( W f h h t 1 + W f x x t + b f )
i t = σ ( W i h h t 1 + W i x x t + b i )
C ˜ t = tanh ( W C h h t 1 + W C x x t + b c )
C t = f t C t 1 + i t × C ˜ t
o t = σ ( W o h h t 1 + W o x x t + b o
h t = o t × tanh ( C t )
where f t represents the forget gate, i t the input gate, and o t the output gate. σ is the sigmoid function, h t 1  the output at the previous moment, xt the input at the current moment, and ht the output at current moment.

3.2.2. GRU

Similar to LSTM, GRU is proposed to solve the problems of long-term memory and gradient in back-propagation, but GRU has a simpler structure. It only contains two gates, a reset gate and an update gate, which reduces the amount of calculation it must do. Its internal structure is shown in Figure 4, and the network structure is as same as LSTM.
The reset gate is used to control the degree of ignoring the state information at the previous moment. The smaller the value of the reset gate, the more it is ignored, and the less of the state information is retained. The update gate is used to control the degree of the previous moment’s state being brought into the current state. Different from LSTM, the output of the GRU’s unit contains only h t 1 , both as the state information of the previous moment to the unit of the next moment and as the input value of the next layer.
r t = σ ( W r h h t 1 + W r x x t + b r )
h ˜ t = ϕ h ( W h h ( r t × h t 1 ) + W h x x t + b h )
z t = σ ( W z h h t 1 + W z x x t + b z )
h t = ( 1 z t ) × h ˜ t + z t × h t 1
where r t represents the reset gate, and z t is the update gate. h ˜ t represents the candidate output value at a current time determined by the reset gate, and ϕ h is a hyperbolic tangent function.
Although the structure of GRU is simpler than that of LSTM, the performance of the two is comparable on many tasks. The fewer parameters of GRU make it easier to converge, but when the dataset is large, LSTM may perform better. Therefore, both are considered in this paper.

3.2.3. BiLSTM

In addition to the above two RNNs, BiLSTM is also selected as one of the methods. The two-direction structure enables the network to obtain complete past and future context information for each point of the input sequence and can obtain better results in some prediction problems that require context information. The internal structure of BiLSTM is shown in Figure 5.
h t = σ ( w 1 x t + w 2 h t 1 )
h t = σ ( w 3 x t + w 4 h t + 1 )
o t = ϕ h ( w 5 h t + w 6 h t )
where h t is the output of the forward-propagation-layer processing unit at current time t , and h t is the output of the back-propagation-layer processing unit.

3.3. Decision-Level Fusion

Considering that the abovementioned methods may appear to have different effects on different datasets, to propose a more applicable model, a decision-level fusion strategy was adopted. Specifically, the method of D–S evidence theory is used.
D–S evidence theory is a theory that deals with the uncertainty that was first proposed by Dempster and further developed by G. Shafer. In D–S evidence theory, the required priori data are more intuitive and easier to obtain than in probabilistic reasoning theory. In addition, D–S evidence theory can synthesize the knowledge or data from different experts or data sources. It has the ability to directly express “uncertain” and “unknown. and these pieces of information are represented in the mass function and retained during the evidence synthesis process. These advantages make the D–S evidence theory widely used [23,24]. The theory is defined, and the synthesis process is detailed as follows.
Letting Ω be a recognition frame (or hypothetical space), then the following is defined.
(1)
Basic probability allocation (BPA)
The BPA in the recognition framework Ω is a function m , called the mass function, and satisfies,
{ m ( ) = 0 A Ω m ( A ) = 1
where A is called focal elements that makes m ( A ) > 0 .
(2)
Belief function
On the recognition framework Ω , the belief function based on m is defined as:
B e l ( A ) = B A m ( B )
(3)
Plausibility function
On the recognition framework Ω , the plausibility function based on m is defined as:
P l ( A ) = B A m ( B )
(4)
Belief interval
In the evidence theory, for a certain hypothesis A in the recognition framework, the BPA is calculated according to the basic probability distribution to calculate the belief function B e l ( A ) and the plausibility function P l ( A ) of the hypothesis to form a belief interval [ B e l ( A ) , P l ( A ) ] , which is used to indicate the degree of confirmation of a certain hypothesis.
(5)
Dempster’s combinational rule
The combinational rules of mass functions are:
( m 1 m 2 m n ) ( A ) = 1 K A 1 A 2 A n m 1 ( A 1 ) m 2 ( A 2 ) m n ( A n )
where K is the normalization constant, calculated by
K = A 1 A n m 1 ( A 1 ) m 2 ( A 2 ) m n ( A n ) = 1 A 1 A n = m 1 ( A 1 ) m 2 ( A 2 ) m n ( A n )
In actual fusion, since the predicted label has only one result and there is no overlap, the element of the recognition framework is equal to the actual category of the dataset in this paper, and the probability result of each network for each sample is the mass function of the network. The fusion structure is shown in Figure 6.

4. Experiments

The experiments in this paper include two parts. In the first part, the RNN variants are used to classify the original time-series data, and the BLS is used to classify the images and evaluate the accuracy separately. In the second part, decision-level fusion is used to fuse the method results in the first part and evaluate and compare the accuracy.
The data used in this article are from the public time-series dataset UCRArchive_2018 [25], which contains 128 sub-datasets contributed by different researchers. The number of samples and sample length included in each sub-dataset are different, but they are all univariate time-series data, and the training and test sets have been divided. The 128 sub-datasets contain a total of 16 types of data, such as sensor data, edge data of objects in the images, simulation data, and motion data of objects. In this paper, a total of seven sub-datasets in four categories are used to conduct experiments to verify the proposed model. Details of the datasets are given in the following subsections.

4.1. BLS and RNN Experiments

4.1.1. BLS Classification Experiment with Encoded Images

In this subsection, both GADF and GASF and RP images are used. During image generation, three image sizes and pixels were fixed. In practice, grayscale images were used for the experiments. The grayscale image reduces the dimension of the input data of the BLS network as well as the amount of calculation relative to the three-channel color image while ensuring recognition accuracy. The BLS network parameters use the same settings as Yang et al. [18]. The number of feature map nodes in each window is 10, there is a total of 10 windows, and the number of feature enhancement nodes is 1500.
As shown in Table 1, using the BLS to classify images of time-series data is effective. However, compared with the GASF method, the overall recognition rate of the GADF method is higher, especially for images that are more difficult to distinguish between classes, such as the image data of SyntheticControl.
As shown in Figure 7, the samples of the two categories on the left-hand side are completely different from the perspective of timing, and the trends are opposite, but they will be very similar after being reversed. From the polar coordinates in the middle of the figure, the mapped time-series data are also basically symmetrical. The difference between the two data points is between π and π , and the sum is between 0 and 2 π . If the cos ( ϕ i + ϕ j ) function is used for calculation for two different angles, the order of addition will not affect the calculation result, giving GASF images a high degree of similarity and making it difficult to classify them accurately. On the contrary, if the sin ( ϕ i ϕ j ) function is used for calculation, the difference value between two different angles is opposite for two orders so that the calculation result is also the opposite. Therefore, the GADF method can better distinguish such data. The results of the improved RP method in this paper are similar to the GADF results. The main reason is that the two images are similar, and both can distinguish the images very well. However, similar to RNN, the BLS cannot effectively distinguish time series with little difference, such as the depiction of similarly shaped leaves in the OSULeaf dataset. The difference in time-series is quite small, which will lead to overly high image similarity.
A characteristic of the BLS network is that it only needs one epoch of calculation to obtain the result, and once the input and network structure are determined, the result is relatively stable, which is completely different from a DL network. The training result of the latter depends on the setting of network parameters and is prone to fluctuation. In addition, another advantage of BLS is that the training time of one epoch is very short. Even for the OSULeaf dataset, with many samples and long data length, the training time is less than 10 s.

4.1.2. RNN Classification Experiments with Time-Series Data

In this experiment, because the data are not particularly large, to ensure accuracy and as little calculation time as possible, all the three RNN variant network structures in this paper have two hidden layers, a fully connected layer, and a SoftMax layer for classification.
The rules of early stopping have been adopted for the three RNN networks. When training DL networks, the best generalization performance is desired; that is, the data must be well fitted. However, usually, because the hyperparameters are not easy to set, especially the training epoch, the problem of overfitting may occur. Although the network performance improves on the training set and the error rate becomes lower, actually, in some moments, its performance on the testing set has begun to deteriorate. One of the methods that is widely used to solve overfitting problems is to set early stopping rules. The performance of the model is calculated on the validation set during training, and when the performance begins to decline, training is stopped so that the problem of overfitting can be avoided. Since there is no additional validation set in the experiment described in this paper, each generation of the model directly uses all of the testing set to test the performance, and the test accuracy is selected as an indicator of early stopping. To prevent the situation of the training being stopped due to unstable shocks at the beginning, another 75 generations is set when the indicator satisfies the stopping conditions to obtain more stable results.
The proposed network has adopted the Dropout setting, which can prevent overfitting and reduce training time. As is well known, when the number of parameters is increasing, the training speed of the model will be affected obviously. With the Dropout strategy, the resulting training time will be greatly reduced by selectively ignoring some hidden-layer neurons in each epoch. Therefore, Dropout is necessary in our framework for the sake of efficiency.
In addition, the SoftMax activation function is used in the multi-classification problem, and the output is turned into the probability format. As a result, the categorical cross-entropy is chosen as the loss indicator. For the network optimizer, the Adam optimizer is used in the proposed framework. Compared with the Stochastic Gradient Descent (SGD) optimizer, Adam does not need the manually selected initial learning rate, and the optimal value can be automatically adjusted during the training process. Moreover, Adam is easy to implement, computationally efficient, and suitable for scenarios with large-scale data and parameters.
As can be seen from Table 2, the GRU and BiLSTM are superior in terms of accuracy. To further demonstrate the efficiency of the two methods, the averaged time consumption of one epoch for different datasets are compared, and the results are listed in Table 3, where it can be clearly seen that GRU takes obviously less time due to its structural superiority. It needs to be pointed out that the training time might vary with different hardware facilities, experimental environments, etc., in real applications. Although LSTM is not as good as the other two networks in terms of performance, its performance on some datasets is still acceptable, so all three networks are considered for later fusion.

4.2. Decision-Level Fusion Experiment and Results Comparison

Considering that different methods exhibit different performances on the same dataset, to ensure that the classification results of images and time series can be reflected in the fusion, a multi-combination fusion method is adopted. At least one result of using time-series data and one of image data is selected for fusion using D–S evidence theory, so there are a total of 13 combinations. The best combination is selected as the final classification result. As shown in Table 4, the best results obtained by fusion are higher than those obtained using a single network in all datasets. Compared with the average accuracy rate, the improvement rate is up to 20.68%. To further verify the performance of the proposed framework, more metrics are introduced to discuss the obtained results. Table 5 shows the results of three evaluation indexes, which are precision rate, recall rate, and F1-score. It can be clearly seen that the proposed model is with the best performance. Among all of the datasets, the averaged precision rate, recall rate, and F1-score are increased by the proposed model with the ratio of 8.5%, 6.82%, and 7.65%, respectively. Thus, the proposed framework is approved to be effective.

5. Conclusions

In this paper, BLS is used to classify the images of time-series data, and three recurrent neural networks, i.e., LSTM, GRU, and BiLSTM, were used to classify the time-series data. The BLS and D–S evidence theories are used to combine multiple decision fusion results to select the highest accuracy rate. The results of experiments prove the effectiveness of the proposed framework.
In image classification, the BLS method can quickly and efficiently classify images with lower complexity. Compared with other deep networks, the BLS method can save a significant amount of training time. In terms of overall time usage, the time from encoding time-series data to images to using the BLS for learning is similar, or even less, than using time-series data and RNN variant networks for classification. However, to better improve the applicability of the model to the data, two features are indispensable. In the direct learning and classification of time-series, the series model of RNN is a very good choice due to its memory of the time relationship of the sequence data. LSTM solves the problem of long-term dependence of a traditional RNN through the control of information by forget, input, and output gates, while GRU simplifies the three gates into a reset gate and an update gate. The two performances are similar in most situations. BiLSTM solves the problem of requiring contextual information. In the method of encoding a time series as an image, the GAF and RP methods can intuitively show the time relationship between the sequence data through the image.
Finally, in decision-level fusion, the D–S evidence theory is considered a strategy that can synthesize the results of different decision-making methods; moreover, it does not need to meet the probability additivity requirements. To further improve the classification accuracy, the use of at least one original time-series dataset and one image data results set is guaranteed in this paper, and multi-combination decision-level fusion is carried out to achieve the purpose of fusing the best model.
In future research, the framework proposed in this paper will continue to be improved to solve the problem of fast and efficient classification of multivariate time series.

Author Contributions

Conceptualization, X.-B.J. and T.S.; methodology, A.Y. and Y.B.; software, A.Y. and J.-L.K.; validation, A.Y.; writing—original draft preparation, A.Y.; writing—review and editing, T.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China (61903009, 62006008), Beijing Municipal Education Commission (KM201810011005, KM201910011010), Young Teacher Research Foundation Project of BTBU (QNJJ2020-26), and Beijing excellent talent training support project for young top-notch team No. 2018000026833TD01.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data supporting reported results can be found in https://www.cs.ucr.edu/~eamonn/time_series_data_2018/, accesseed on 20 September 2019.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, S.B. Asymptotic Tracking Control for Nonaffine Systems with Disturbances. IEEE Trans. Circuits Syst. II Express Briefs 2021. [Google Scholar] [CrossRef]
  2. Wang, S.; Na, J.; Chen, Q. Adaptive Predefined Performance Sliding Mode Control of Motor Driving Systems with Disturbances. IEEE Trans. Energy Convers. 2020. [Google Scholar] [CrossRef]
  3. Nour, M.; Cömert, Z.; Polat, K. A novel medical diagnosis model for COVID-19 infection detection based on deep features and Bayesian optimization. Appl. Soft Comp. 2020, 97, 106580. [Google Scholar] [CrossRef] [PubMed]
  4. Jin, X.-B.; Yu, X.-H.; Su, T.-L.; Yang, D.-N.; Bai, Y.-T.; Kong, J.-L.; Wang, L. Distributed Deep Fusion Predictor for a Multi-Sensor System Based on Causality Entropy. Entropy 2021, 23, 219. [Google Scholar] [CrossRef] [PubMed]
  5. Huang, J.; Wei, Y.Q.; Yi, J.; Liu, M.D. An Improved kNN Based on Class Contribution and Feature Weighting. In Proceedings of the 2018 10th International Conference on Measuring Technology and Mechatronics Automation (ICMTMA), Changsha, China, 10–11 February 2018; pp. 313–316. [Google Scholar]
  6. Hu, X.Y.; Mo, S.R.; Peng, D.S.; Shen, F.; Luo, C.; Qu, S.D. Automatic Activity Classification Based on Human Body Kinematics and Dynamic Time Wrapping. In Proceedings of the 2018 IEEE 23rd International Conference on Digital Signal Processing (DSP), Shanghai, China, 19–21 November 2018; pp. 1–5. [Google Scholar]
  7. Do, C.; Douzal-Chouakria, A.; Marié, S.; Rombaut, M. Multiple Metric Learning for Large Margin kNN Classification of Time Series. In Proceedings of the 23rd European Signal Processing Conference (EUSIPCO), Nice, France, 31 August–4 September 2015; pp. 2346–2350. [Google Scholar]
  8. Kurt, İ.; Ulukaya, S.; Erdem, O. Classification of Parkinson’s Disease Using Dynamic Time Warping. In Proceedings of the 27th Telecommunications Forum (TELFOR), Belgrade, Serbia, 26–27 November 2019; pp. 1–4. [Google Scholar]
  9. Xu, Z.Q.; Zhao, W.D.; Li, A.J.; Peng, L.X. Research on the Hidden Markov Model in Hot Topics of the Internet. In Proceedings of the 2018 IEEE 18th International Conference on Communication Technology (ICCT), Chongqing, China, 8–11 October 2018; pp. 1315–1318. [Google Scholar]
  10. Cherkassky, V. The Nature of Statistical Learning Theory. IEEE Trans. Neural Netw. Learn. Syst. 1997, 8, 1564. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Wang, M.; Abdelfattah, S.; Moustafa, N.; Hu, J. Deep Gaussian Mixture-Hidden Markov Model for Classification of EEG Signals. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 2, 278–287. [Google Scholar] [CrossRef]
  12. Alickovic, E.; Subasi, A. Ensemble SVM Method for Automatic Sleep Stage Classification. IEEE Trans. Instrum. Meas. 2018, 67, 1258–1265. [Google Scholar] [CrossRef] [Green Version]
  13. Dutta, K.K. Multi-class Time Series Classification of EEG Signals with Recurrent Neural Networks. In Proceedings of the 2019 9th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India, 10–11 January 2019; pp. 337–341. [Google Scholar]
  14. Jin, X.-B.; RobertJeremiah, R.J.; Su, T.-L.; Bai, Y.-T.; Kong, J.-L. The New Trend of State Estimation: From Model-Driven to Hybrid-Driven Methods. Sensors 2021, 21, 2085. [Google Scholar] [CrossRef] [PubMed]
  15. Kong, J.; Wang, H.; Wang, X.; Jin, X.-B.; Fang, X.; Lin, S. Multi-stream hybrid architecture based on cross-level fusion strategy for fine-grained crop species recognition in precision agriculture. Comput. Electron. Agric. 2021, 185, 106134. [Google Scholar] [CrossRef]
  16. Chen, C.L.P.; Wan, J.Z. A Rapid Learning and Dynamic Stepwise Updating Algorithm for Flat Neural Networks and the Application to Time-series Prediction. IEEE Trans. Syst. Man Cybern. Part B 1999, 29, 62–72. [Google Scholar] [CrossRef] [PubMed]
  17. Chen, C.L.P.; Liu, Z. Broad Learning System: An Effective and Efficient Incremental Learning System without the Need for Deep Architecture. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 10–24. [Google Scholar] [CrossRef] [PubMed]
  18. Yang, A.Q.; Yu, X.H.; Su, T.L.; Jin, X.B.; Kong, J.L. Broad Learning System for Human Activity Recognition Using Sensor Data. Int. J. Comput. Appl. Technol. 2019, 61, 259–264. [Google Scholar] [CrossRef]
  19. Wang, Z.G.; Oates, T. Encoding Time Series as Images for Visual Inspection and Classification Using Tiled Convolutional Neural Networks. In Proceedings of the Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–26 January 2015. [Google Scholar]
  20. Hatami, N.; Gavet, Y.; Debayle, J. Classification of Time-Series Images Using Deep Convolutional Neural Networks. In Proceedings of the 2017 the 10th International Conference on Machine Vision (ICMV 2017), ICMV Committees, Vienne, Austria, 13–15 November 2017. [Google Scholar]
  21. Karimi-Bidhendi, S.; Munshi, F.; Munshi, A. Scalable Classification of Univariate and Multivariate Time Series. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 1598–1605. [Google Scholar]
  22. Eckmann, J.P.; Kamphorst, S.O.; Ruelle, D. Recurrence Plots of Dynamical Systems. Europhys. Lett. 1987, 4, 973–977. [Google Scholar] [CrossRef] [Green Version]
  23. Chen, B.C.; Tao, X.; Yang, M.R.; Yu, C.; Pan, W.M.; Leung, V.V.M. A Saliency Map Fusion Method Based on Weighted DS Evidence Theory. IEEE Access 2018, 6, 27346–27355. [Google Scholar] [CrossRef]
  24. Fei, L.; Xia, J.; Feng, Y.; Liu, L. An ELECTRE-Based Multiple Criteria Decision Making Method for Supplier Selection Using Dempster-Shafer Theory. IEEE Access 2019, 7, 84701–84716. [Google Scholar] [CrossRef]
  25. Dau, H.A.; Keogh, E.; Kamgar, K.; Yeh, C.C.M.; Zhu, Y.; Gharghabi, S.; Ratanamahatana, C.A.; Chen, Y.P.; Hu, B.; Bejum, N.; et al. The UCR Time Series Classification Archive. Available online: https://www.cs.ucr.edu/~eamonn/time_series_data_2018/ (accessed on 20 September 2019).
Figure 1. Framework.
Figure 1. Framework.
Sensors 21 04391 g001
Figure 2. The conversion of a time series using RP and GAF methods.
Figure 2. The conversion of a time series using RP and GAF methods.
Sensors 21 04391 g002
Figure 3. BLS network.
Figure 3. BLS network.
Sensors 21 04391 g003
Figure 4. LSTM and GRU networks.
Figure 4. LSTM and GRU networks.
Sensors 21 04391 g004
Figure 5. BiLSTM networks.
Figure 5. BiLSTM networks.
Sensors 21 04391 g005
Figure 6. Fusion structure.
Figure 6. Fusion structure.
Sensors 21 04391 g006
Figure 7. GADF example of stronger image discrimination.
Figure 7. GADF example of stronger image discrimination.
Sensors 21 04391 g007
Table 1. The classification accuracy of images from different datasets using the BLS network.
Table 1. The classification accuracy of images from different datasets using the BLS network.
DomainDatasetTrainTestLengthBLS + GASFBLS + GADFBLS + RP
MotionGunPoint5015015095.3397.3393.33
SensorLightning7707331961.6465.7563.01
Trace10010027510010096
ImageOSU Leaf20024242753.3157.43858.68
SimulatedSyntheticControl3003006061.3397.3398
CBF3090012894.2295.1197.22
UMD3614415068.7581.9497.22
Table 2. The classification accuracy of different datasets using RNN series network.
Table 2. The classification accuracy of different datasets using RNN series network.
DomainDatasetTrainTestLengthLSTMGRUBiLSTM
MotionGunPoint5015015097.339697.33
SensorLightning7707331958.972.660.27
Trace1001002755210071
ImageOSULeaf20024242757.4459.9259.92
SimulatedSyntheticControl30030060989998
CBF3090012898.7899.1198
UMD3614415079.8698.61100
Table 3. Averaged time consumption of one epoch for different datasets (Unit: second).
Table 3. Averaged time consumption of one epoch for different datasets (Unit: second).
Networks.GunPointLightning7TraceOSULeafSyntheticControlCBFUMD
GRU1.134.744.1726.560.361.521.05
BiLSTM1.88.27.6847.81.532.851.73
Table 4. A comparison of accuracy before and after fusion.
Table 4. A comparison of accuracy before and after fusion.
DatasetLSTM (%)GRU (%)BiLSTM (%)BLS + GADF (%)BLS + RP (%)Best D–S (%)Improvement (%)
GunPoint97.339697.3397.3393.33981.77
Lightning758.972.660.2765.7563.0180.8220.68
Trace52100711009610016.20
OSULeaf57.4459.9259.9257.43858.6866.5311.80
SyntheticControl98999897.339899.671.61
CBF98.7899.119895.1197.2299.672.03
UMD79.8698.6110081.9497.221008.47
Table 5. Evaluation indexes of different models.
Table 5. Evaluation indexes of different models.
NetworksGunPointLightning7TraceOSULeafSyntheticControlCBFUMD
LSTMPrecision0.97440.59080.26210.56290.98070.98790.8197
Recall0.97370.54860.50.57830.980.98780.7986
F1-score0.97040.56890.38110.57050.98030.98790.809
GRUPrecision0.96020.74610.60730.99010.99130.9867
Recall0.96020.709610.60940.990.99110.9861
F1-score0.96020.727410.60830.99010.99120.9864
BiLSTMPrecision0.97440.62390.75450.58820.98030.98021
Recall0.97370.43130.72670.57810.980.98011
F1-score0.9740.510.74030.58310.98020.98021
BLS + GADFPrecision0.97350.629410.60180.97470.95410.8167
Recall0.97350.604810.58970.97330.95140.8194
F1-score0.97350.616810.59570.97400.95270.8178
BLS + RPPrecision0.93650.54560.95650.60880.98090.97270.9723
Recall0.93280.56620.96430.59430.98000.97240.9722
F1-score0.93460.55570.96040.60150.98040.97250.9723
Best D–SPrecision0.98050.795010.68570.99670.99441
Recall0.98030.670910.67020.99670.99451
F1-score0.98040.727710.67790.99670.99451
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jin, X.-B.; Yang, A.; Su, T.; Kong, J.-L.; Bai, Y. Multi-Channel Fusion Classification Method Based on Time-Series Data. Sensors 2021, 21, 4391. https://doi.org/10.3390/s21134391

AMA Style

Jin X-B, Yang A, Su T, Kong J-L, Bai Y. Multi-Channel Fusion Classification Method Based on Time-Series Data. Sensors. 2021; 21(13):4391. https://doi.org/10.3390/s21134391

Chicago/Turabian Style

Jin, Xue-Bo, Aiqiang Yang, Tingli Su, Jian-Lei Kong, and Yuting Bai. 2021. "Multi-Channel Fusion Classification Method Based on Time-Series Data" Sensors 21, no. 13: 4391. https://doi.org/10.3390/s21134391

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop