Next Article in Journal
Organic Scintillator-Fibre Sensors for Proton Therapy Dosimetry: SCSF-3HF and EJ-260
Previous Article in Journal
Experimental Analysis to Detect Corona COVID-19 Virus Symptoms in Male Patients through Breath Pattern Using Machine Learning Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Dendritic Cell-Negative Selection Method for Earthquake Prediction

1
School of Computer Science, Hubei University of Technology, Wuhan 430068, China
2
School of Computer Science, Wuhan University, Wuhan 430072, China
*
Authors to whom correspondence should be addressed.
Electronics 2023, 12(1), 9; https://doi.org/10.3390/electronics12010009
Submission received: 10 November 2022 / Revised: 9 December 2022 / Accepted: 14 December 2022 / Published: 20 December 2022
(This article belongs to the Topic Advanced Systems Engineering: Theory and Applications)

Abstract

:
Earthquake prediction (EQP) is an extremely difficult task, which has been overcome by adopting various technologies, with no further transformation so far. The negative selection algorithm (NSA) is an artificial intelligence method based on the biological immune system. It is widely used in anomaly detection due to its advantages of requiring little normal data to detect anomalies, including historical seismic-events-based EQP. However, NSA can suffer from the undesirable effect of data drift, resulting in outdated normal patterns learned from data. To tackle this problem, the data changes must be detected and processed, stimulating fast algorithmic adaptation strategies. This study proposes a dendritic cell algorithm (DCA)-based adaptive learning method for drift detection and negative selection algorithm (DC-NSA) that dynamically adapts to new input data. First, this study adopts the Gutenberg–Richter (GR) law and other earthquake distribution laws to preprocess input data. Then, the NSA is employed for EQP, and then, the dendritic cell algorithm (DCA) is employed to detect changes to trigger gradient descent strategies and update the self-set in NSA. Finally, the proposed approach is implemented to predict the earthquakes of MW > 5 in Sichuan and the surroundings during the next month. The experimental results demonstrate that our proposed DC-NSA is superior to the existing state-of-the-art EQP approaches.

1. Introduction

EQP aims at predicting the specific location, magnitude and time of future earthquake events [1]. Through long-term observation, research and accumulation of experience by experts, it is found that there is an inherent relationship between the three elements of earthquakes (earthquake occurrence time, magnitude, and epicenter location) and characteristic indicators [2]. Prediction of changes using historical seismic characteristic indicators is a commonly used method for EQP.
With the development of artificial intelligence, researchers have tried to apply related algorithms to EQP, and achieved good prediction results. Large earthquakes is a small probability event with a long time interval, and the number of large earthquakes observed by modern observation equipment is relatively few. Therefore, the lack of data affects the accuracy of EQP using historical seismic data [1].
The NSA [3], which belongs to an artificial immune algorithm, simulates the principle of self–non-self identification in the biological immune system. The algorithm mimics the process of the negative selection of immature T lymphocytes in the organism during the maturation process. First, determine the self-set; then randomly generate detectors, by deleting the detectors that recognize the self, and keep the detectors that recognize the non-self. Its advantage is that it does not require prior knowledge, and only needs a limited amount of self-data as a training set to generate a large number of non-self detectors for detecting the non-self [3]. It is often used in binary classification problems, for it can identify self and non-self. In this study, predicting the occurrence of earthquakes of moment magnitude (MW) 5.0 and above can be regarded as a binary classification problem in essence [4], and the NSA can be used to identify the occurrence of earthquakes (non-self) and the non-occurrence of earthquakes (self). As there is no need for non-self datasets (earthquakes of MW 5.0 and above) in the negative selection training process, the low prediction accuracy problem caused by the lack of large earthquake data in traditional machine learning methods can be solved.
In this study, the NSA of computer immunology is introduced to establish the EQP model. Using the self-non-self identification principle of negative selection, earthquakes with a predefined threshold and above magnitudes that did not occur are regarded as self, and earthquakes with a predefined threshold and above magnitudes are regarded as non-self. Since only the self data are used as the training set, the influence of the lack of large earthquake data on the training effect is reduced, and the purpose of improving the prediction accuracy of large earthquakes is achieved. Specifically, the gradient descent that encompasses DCA for data drift identification is used for the optimization of NSA. The DCA is executed in an online fashion to detect changes from the input data. A detected data drift will stimulate the optimization process. In the experimental part of this study, the MW 5.0 is used as our prediction threshold.
The main contributions of this study are described as follows:
  • A NSA model with the ability of adaptive learning is proposed, and it can adapt to changes in the environment and context.
  • DCA detects data drifts in the context of trigger-adaptation strategies, namely, the gradient descent is used to optimize the radius parameter of NSA.
  • The proposed DC-NSA is implemented to the historical seismic events in Sichuan and the surrounding areas.
The remainder of this study is summarized as follows: Section 2 reviews the related works; the proposed DC-NSA framework is described in Section 3; the experimentation part, including baselines and comparison analysis, are demonstrated in Section 4; and Section 5 introduces the conclusion and future works.

2. Related Works

Various machine learning methods have been implemented in EQP using historical seismic events, including statistic methods, ANN-based approaches, deep learning, and artificial immune methods [5,6].
Several rule-based approaches have been proposed to implement EQP tasks. For instance, Dehbozorgi and Farokhi [7] first introduced neuro-fuzzy to predict short-term earthquake via historical seismic events. The time, location, seismic magnitude, depth, statistical, and entropy parameters, are adopted to predict whether an earthquake will occur in the following five minutes. The association rule mining is applied to predict a subsequent earthquake with reference to the historical seismic events [8]. A spatial analysis of magnitude distribution for EQP using ANFIS based on automatic clustering is proposed to predict earthquakes of magnitudes higher than MW 5 in Indonesia [9]. Rule-based approaches are popular due to their ease of use and flexibility.
Many studies, such as [10,11,12], have been proposed to deal with EQP. Of these, artificial neural networks (ANNs) are the widely used approaches due to their ability to engage in self-learning and handle complex problems. However, it suffers from the lack of training data when implemented for EQPs. Shi et al. [13] first introduced ANN in the EQP context and established the relationships between magnitude and earthquake epicentral intensity; nevertheless, the method achieves poor performance. Many research works have followed this method using various ANNs. For instance, a support vector regressor and hybrid neural network (SVR-HNN) are adopted to predict earthquakes in [14]; this approach uses maximum relevance and minimum redundancy criteria for the relevant indicators’ extraction. The authors provide sixty seismic features for EQP and achieve encouraging prediction results. However, this work does not analyze seismic-related and non-seismic-related anomalies.
There are also many shallow machine learning works that focus on EQP. For example, in [15], a method to discover clustering-based patterns and predict medium-large earthquakes is proposed. This approach adopts maximum likelihood to estimate the b-value of seismic data. The authors confirm that the b-value can be regarded as an earthquake precursor, and K-means achieves good performance. However, it only considers the b-value as a seismic precursor, and thus cannot truly reflect the complexity of an earthquake and affect the prediction performance. In [16], the principal-component-analysis-based random forest (PCA-RF) was introduced to execute data dimension reduction and generate new datasets to generalize existing prediction models, and the experimental results demonstrate that the average accuracies of these approaches have been improved. Unfortunately, the differences in geological structure hinder their universality. Asim et al. [17] adopted support vector machine (SVM) and random forest to predict earthquake activities in Cyprus, while further proving that random forest is most suitable for magnitude thresholds of MW 3.0 and MW 3.5. Notably, however, random forest produces overfitting.
Several deep learning approaches have been proposed to implement EQP tasks. For instance, DeVries et al. [18] introduced a deep-learning method to determine a static stress-based criterion that can predict aftershock locations without presupposing the fault orientation. Moreover, it provides improved aftershock prediction location and identifies physical quantities that control seismic triggering when the earthquake cycle is active; however, it requires a large amount of training samples. Huang et al. [19] adopted CNN for continuous EQP using historical seismic events in Taiwan, and identified its temporal pattern, which may be useful for further EQP. However, this research conducts no analysis regarding seismic indicators. Wang et al. [20] introduced LSTM to learn the spatio-temporal relationships among earthquakes and further proved its the robustness and effectiveness. This approach can be used for EQP, even in areas without seismic sensors; however, it is computationally expensive and time-consuming.
Various artificial immune approaches have been implemented in the EQP context. For instance, in [21], NSA is adopted to reduce the impact of inadequate earthquake data on the performance in training phase. First, eight seismic indicators proposed in [11] served as the input of NSA. If an earthquake with a magnitude (MW ≥ 4.5) does not occur, it is regarded as “self”; otherwise, it is regarded as “non-self”. A mature detector is then generated to detect anomalous test instances. While this approach can quickly detect earthquakes, the self-detectors are difficult to define, and the deletion of matching detectors in the detector generation phase will lead to low algorithmic efficiency. In [22], DCA is introduced for EQP. First, PCA is adopted to map these indicators to the safe signal (SS), pathogen-associated molecular pattern (PAMP) and danger signal (DS) of DCA; here, PAMP denotes the existence of anomalous indicators (indicating strong seismic activity), SS indicates that the possibility of normal is relatively high (that is, the seismic activity is weak), and DS shows that the possibility of an earthquake is high. DC then randomly samples antigens and signals to produce cumulative co-stimulatory molecules (csm), a semi-mature signal (semi), and a mature signal (mat). When the cumulative csm exceeds a given migration threshold, DC begins to migrate. If the cumulative semi exceeds the cumulative mat, the DC is semi-mature; otherwise, it is mature. Finally, the number of times each antigen is presented to be normal and anomalous is calculated, and the degree of abnormality of the antigen is evaluated by the calculating mature context antigen value (MCAV). If MCAV exceeds an anomalous threshold, the antigen is anomalous (an earthquake occurs); otherwise, it is normal (an earthquake does not occur). In [23], a Haskell-based deterministic DCA (EQP-hDCA) was presented to predict the magnitude of earthquakes in Sichuan and surroundings with magnitudes greater than MW 4.5 in the following month. While the DCA-based EQP obtains good performance, it is affected by a serious false alarm rate, while dealing with the application, including frequent data-type switching.
To sum up, the selection of the EQP model is very important to the prediction performance. This study proposes a novel EQP method; we select the eight indicators used in literature [2] as eigenvectors, and use negative selection to establish an EQP model. Moreover, DCA is used for the optimization of NSA. Since the detector generation process of the NSA only needs to use the self dataset, it can be used to reduce the problem of low prediction accuracy caused by the lack of large earthquake data in methods such as neural networks. Then, the DCA detects data drifts in input data to trigger gradient descent adaptation strategies, to improve the prediction accuracy.

3. The Proposed DC-NSA Earthquake Prediction Approach

The DC-NSA architecture for EQP is described in Figure 1, including three parts: preprocessing, NSA prediction and DC-based adaptive learning process. Firstly, the seismic event set is described as E = { E 1 , , E i , E N } , which is the data source of seismic indicator matrix F, as explained in Table 1. According to [24], L a b e l = 1 indicates that the maximum magnitude for events occurring within the prediction horizon is larger than or equal to the threshold M m e a n , and L a b e l = 0 otherwise. The input of DC-NSA is matrix F. Then, the corresponding DCA model executes a signal acquisition operation to sensor changes or drifts in the contextual data, and initializes the parameters of the parameters of the gradient descent. Meanwhile, the obtained signal matrix in DCA model serves as the input of NSA to execute an EQP task. Then, with the recognition of changes or drifts in contextual data, gradient descent is implemented to optimize the radius r s of NSA. The output of NSA-DCA is the prediction result; 1 indicates an earthquake of magnitude equal to or greater than M s occurs, and 0 otherwise.

3.1. Seismic Indicators

The seismic indicators are calculated according to GR law, the distribution of earthquakes magnitude, and the latest EQP studies. In this study, the indicator matrix is defined as F = { F 1 , , F i , F N } , 0 < i < N , where N is the number of seismic events, while F i = { b i , η i , Δ M i , T M s i , μ M s i , c M s i , d E i 1 / 2 , M m e a n i } [11]. For a seismic event i, the indicator b i is the b-value, η i is the mean square deviation, Δ M i denotes the magnitude difference between the maximum magnitude and the maximum expected magnitude observed according to GR law, T M s i is the time elapsed over the last N i events with a magnitude greater than a predefined threshold ( M s ), μ M s i is the mean time between the N last M s -characteristic events, c M s i is the coefficient of variation of the mean time between the N last M s -characteristic events ( μ M s i ), d E i 1 / 2 is the rate of the square root of released seismic energy, and M m e a n i is the mean of the Richter magnitudes of the last N events. Table 1 presents the formulas used to calculate these values. Here, a i denotes the cumulative frequency of earthquakes above zero, M i denotes the Richter magnitude, n is the total number of earthquake events, t i denotes the observed elapsed time between characteristic events ( M i ), and n c is the total number of characteristic events. The first three indicators and the difference Δ M i are related to GR law, while T M s i , d E i 1 / 2 , M m e a n i are unrelated to the assumed temporal distribution of seismic magnitude; moreover, μ M s i , c M s i are related to the distribution of the characteristic temporal seismic magnitude.

3.2. Negative Selection Algorithm

Negative selection aims at generating tolerance to self cells, and hence, the immune system has the ability to detect unknown antigens and not to react to self-cells [3]. The input of the NSA is self set, where each self is represented by an n-dimensional vector. The detectors (antibodies) generated by the algorithm cover the non-self space. It mainly uses an n-dimensional vector to define the detector, and uses a real value to describe the detection radius of the detector. A detector can be thought of as a hypersphere. Whether the detector detects the antigen depends on the distance between the detector and the antigen (calculated using the Euclidean distance formula) and the detector radius. If the distance is less than the self-radius, which means that the two hyperspheres intersect, which means that the detector can identify self, then the detector is eliminated; otherwise, the distance between the detector and self is taken as the radius. If the mature detectors have different radii, a large range of non-self spaces can be covered by very few discriminators; meanwhile, for less non-self spaces around the self, detectors with smaller radii can be used to cover them. The mature detector generation algorithm flow is illustrated in Algorithm 1.
Algorithm 1 NSA
Input: self S
Output: classification matrix
  1: Initialize the number of detectors n u m , self radius r s , detectors set D, and maximum distance d i s t a n c e
  2: while the number of detectors < n u m do
  3:    Generates n-dimensional vectors d randomly
  4:    for each antigen s in training set S do
  5:      Calculate the Euclidean distance d i s between the random detector d and the self
  6:      if  d i s < d i s t a n c e then
  7:          d i s t a n c e = d i s
  8:      end if
  9:    end for
  10:   if r s < d i s t a n c e then
  11:       r d = d i s t a n c e r s
  12:      Add the vector < d , r d > to the mature detector set D
  13:    end if
  14: end while

3.3. Dendritic Cell Algorithm-Based NSA

In the biology immune system, dendritic cells (DCs) can identify potentially damaging foreign bodies, namely, it has strong classification abilities. Therefore, the DCA was proposed and implemented as an intrusion detection approach [25].
DCA draws on the antigen present process of DCs. DCs are able to collect and process antigens, which are molecules that stimulate immune responses, as they are in the mature state. In the immature state (iDCs), the DCs can collect diverse immune signals, including pathogen-associated molecular pattern signals (PSs), danger signals (DSs), and safety signals (SSs). The iDC changes to a semi-mature state (smDC) or mature state (mDC) according to the concentration of immune signals received. More precisely, smDC contains more SS than PS and DS, and can be declared a security context. On the contrary, mDC exposed to a large number of PS and DS will be considered a danger environment; therefore, the antigen is eliminated. Another output is the costimulatory molecule (CSM). It executes DC migration when a predefined threshold γ M T is reached. Algorithm 2 shows the structure of DCA.
The anomaly metric characterizing the signal is in accordance with the distance metric, which simply calculates the difference between SS and DS in the context parameter values of time t and t 1 . The Euclidean distance between all contextual parameters at time t and t 1 is applied for P S . Given a dataset of k contextual features { B 1 , , B k } , where each feature variable B i can obtain a value from its own support χ i , and n contextual feature instances, b t = ( b 1 , , b k ) χ 1 × χ 2 × × χ k , with t { 1 , , n } , S S , D S and P S signals are calculated as
S t = S t S t 1
D t = D t D t 1
P t = | | b t b t 1 | | F
where | | · | | F is the Frobenius norm, and b t is the Z-score normalized contextual feature vectors at a t . After determining when the DC becomes smDC or the lifetime duration of mDC according to the concentration of its PS, DS and SS signals, check the DC, depending on γ M T . The checking process can be described as
D C s c o r e i = w S S t + w D S t + w P P t w S + w D + w P
where w S , w D , and w P are the weights for the S S , D S , and P S signals, respectively. S t , D t , and P t are the signal values at time t computed based on Equations (1)–(3), respectively.
When the migration threshold γ M T is reached, the DC no longer acknowledges the new instance to process, classifies the antigen and removes the DC from the population. Then it creates a new DC to replace it. A time window T W is adopted to account for temporality relative to the DC state. Checks the fixed segment | t T W | of a DC of size T W , so the resulting value can consider previous values. The γ F T is a failure threshold, and it is finally possible to estimate whether a segment of size | t T W | at time t indicates a change, which in this case, corresponds to a context change, and then the gradient descent should be activated. Assuming that at least one contextual feature is set by the DC as a change at time t, the entire system has to undergo the data drift:
r t = c h a n g e , i f i = 1 k D C i | t T W | > 1 , n o r m a l , o t h e r w i s e .
where r t r = ( r 1 , , r n ) . When r t = c h a n g e , the NSA model corresponds to the current system state and must adapt to the new situation. Table 2 shows examples of possible values for three DCA parameters as part of the experimental setup for real test scenarios discussed later.
Algorithm 2 DCA
  • Input: seismic indicators F.
  • Output: antigen context vectors.
  • for each antigen poll do
  •     S t , D t , and P t calculation by Equations (1)–(3);
  • end for
  • for each DC do
  •    antigen adaptation;
  •    signal fusion calculation by Equation (4) and Table 2;
  •    if C C S M > migration threshold then
  •      antigen adaptation;
  •    end if
  •    for the DC do
  •      if SEMI > γ M T then
  •         DCContext = SEMI;
  •      else
  •         DCContext = MAT;
  •      end if
  •    end for
  • end for
  • for each antigen do
  •    if MCAV > γ F T then
  •      antigen = anomalous;
  •      gradient descent;
  •    else
  •      antigen = normal;
  •    end if
  • end for

4. Experimentation

In this section, the proposed DC-NSA will be validated and compared to various machine learning approaches: namely, DCA [22], NSA [21], PCA-RF [16], back propagation neural network (BPNN) [11], RNN [2], probabilistic neural network (PNN) [12], EQP-hDCA [23], LSTM [19] and SVR-HNN [14].

4.1. The Dataset and Performance Measurement

Information about earthquakes from Gansu (DS1), Qinghai (DS2), Sichuan (DS3) and Yunnan (DS4) from 1 January 1990 to 13 October 2018 was obtained from the China National Earthquake Science Data Center [26]. As it is of great significance to predict earthquakes with a magnitude greater than a threshold, we selected 12,040 data points with magnitudes greater than MW 3.0 as our data source. Table 3 presents the distributions of each extracted dataset. The mean is the average of the magnitudes, and SD is the standard deviation of the magnitudes of each dataset.
Table 4 lists the calculated eight seismicity indicators of Sichuan province according to Table 1, the N is set to 100 [2], and the M s is set as MW   5.0 .
Through the seismic indicators calculation method described in Section 3, all indicator matrices F of DS1, DS2, DS3 and DS4 are calculated, and the box plot is drawn as shown in Figure 2. It can be seen from the figure that there are outliers in the original seismic data and the calculated seismic indicators.

4.2. The Prototype Implementation

In our experiments, the input signals are normalized between 0 and 1; therefore, the DC weight matrix is set according to Table 2; the M s is MW 5.0. Other parameters are as follows: γ M T = 0.5 , γ F T = 0.4 , the learning rate, epochs, and batch size are set to 0.002, 1, and 10, respectively. DCA and EQP-hDCA use PCA as the signal mapping method, create 100 DCs for the pool, and randomly select 10 DCs to sample each antigen. NSA uses the Euclidean distance to measure the similarity between antigen and antibody, and the self radius is set to 0.02. For PCA-RF, the parameters are set as maxDepth = 0, numFeatures = 0, numTrees = 10. The BPNN in our experiment has three layers: input layer, hidden layer and output layer. The learning rate indicator is set to 0.0003, and the number of training periods is set to 100 periods. The weights in the multi-layer perceptron were learned using standard gradient descent. The RNN uses sigmoid as the activation function. The batch size is set to 128, the learning rate indicator is set to 0.001, and the number of training time periods is set to 100 time periods. The weights in the multi-layer perceptron are set for learning using the randomly optimized A method (Adam). The PNN in our experiment has an embedded layer, a product layer and three hidden layers. In LSTM, the learning rate is set to 0.001, and momentum is used to accelerate model convergence, and is set to 0.9. We use 0.0005 weight attenuation, and the abscission rate of the two abscission layers is 0.5. Other parameter settings of the comparison methods can be found in the relevant literature.

4.3. Verification Indicators

The EQP according to historical seismic events is a classification task. Therefore, the following statistical indicators are used to assess the results of the proposed and comparison algorithms: false acceptance rate (FAR), Matthews correlation coefficient (MCC), R-score (R), accuracy (ACC), predictive positive value (PPV), negative predictive value (NPV), recall rate (Rn), specificity (S), and area under the curve (AUC). Apparently, the averages (Avgs) of PPV, NPV, Rn and S are calculated to provide a overall prediction quality measure [11].

4.4. Results Analysis and Comparison

The experimental results of DS1, DS2, DS3 and DS4 are demonstrated in this section. The results of each domain are analyzed separately. Table 5, Table 6, Table 7 and Table 8 show the results of DS1 to DS4, respectively. In all tables, the number in bold indicates the best result among all the methods.
There are only 1132 instances in DS1 area. With reference to DS1 (see Table 5), when using performance evaluation indicators to evaluate all algorithms, the performance is not very good. AUC is only a little higher than 0.5 (random classification results), and R is also poor. However, in such a case, we found that many indicators of DC-NSA are better than NSA. At the same time, in most evaluation indicators, the results are not as good as those of DCA, and are similar to those of DCA in R. Therefore, the results of DC-NSA are not competitive on this dataset.
The domain DS2 has a higher SD, and Table 6 summarizes the results of DS2. It is worth noting that compared with other methods, the predicted FAR of DC-NSA is limited, while the value of R is 1.00, which is a very ideal feature of EQP method. Therefore, the higher SD may achieve higher performance.
There are 5433 instances in DS1 area. The specific results of DS3 are shown in Table 7. The difference among Rn, MCC and R between DC-NSA and other algorithms is very obvious, with the difference exceeding 0.18 units. The difference between the second-best classifier of Rn (EQP hDCA) and the second-best classifier of MCC (LSTM) exceeds 0.28 units. Compared with the second-best classifier of R (EQP hDCA), the difference exceeds 0.27 units. Therefore, the higher SD may achieve higher performance.
Table 8 shows the results of DS4. DC-NSA provides better results on most evaluation indicators, with the smallest FAR and a R value of 0.96. When the other nine algorithms predict DS4, the overall difference between DC-NSA and EQP hDCA is not particularly obvious. In the aspect of index R (the key index of EQP algorithm), NSA is obviously weaker than other algorithms.
From a joint analysis of the above-mentioned tables, since DC-NSA achieves the best results on most indicators, such as R and Avg, we can conclude that DC-NSA is the most suitable classifier for these datasets. Moreover, it can be found that the datasets with higher SD or lager number of instances may achieve higher performance.

5. Conclusions and Future Work

This paper described an EQP method by integrating DCA and NSA. The main contribution of this paper is that it presents a novel DC-NSA approach to make a more suitable EQP model. Firstly, DC-NSA preprocesses earthquake indicators using the GR law and other earthquake magnitude distribution techniques, and obtains a seismic indicator F. Further, the corresponding DCA model executes a signal acquisition operation to sensor changes or drifts in the contextual data, and initializes the parameters of the parameters of the gradient descent. Meanwhile, the obtained signal matrix in DCA model serves as the input of NSA to execute an EQP task. Then, with the recognition of changes or drifts in contextual data, gradient descent is implemented to optimize the radius r s of NSA. Finally, this study used the historical seismic events in Sichuan and surroundings as our experimental data, and compared the proposed approach with DCA, NSA, PCA-RF, BPNN, RNN, PNN, EQP-hDCA, LSTM, and SVR-HNN, using the PPV, NPV, Rn, S, FAR, MCC, AUC, Avg, and R as comparison criteria. The experimental results demonstrate that our proposed DC-NSA is superior to the compared approaches.
In this study, the proposed method is derived from the function of T cell and DCs in the immune system; unlike traditional EQP methods, which lacks adaptively, the proposed approach implements the change adjustments based on the input sample. Meanwhile, the DC-NSA can handle the problem that NSA suffers from the normal patterns learned from data that have become obsolete.
However, this study only focuses on the historical seismic events, neglecting the observed data, which are also important for EQP. Therefore, we will adopt the observed data with much more instances than our experimental data. Moreover, future work will focus on enriching NSA and applying it to other earthquake-prone areas. We will further test NSA to improve the AIS detection performance through subsequent experiments.

Author Contributions

Validation, W.L.; Investigation, Z.M. and J.C.; Writing—original draft, W.Z.; Writing—review & editing, Z.Y. and Q.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NSFC grant number 61877045 and 62202147, Key projects of Scientific Research Project of Hubei Provincial Department of Education grant number D20191406.

Data Availability Statement

Not applicable.

Acknowledgments

The authors want to thank NSFC- http://www.nsfc.gov.cn/ (accessed on 10 November 2022) for the support through Grants Number 61877045 and 62202147, Key projects of Scientific Research Project of Hubei Provincial Department of Education for their support through Grant Number D20191406.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yuntai, C. Earthquake prediction: Retrospect and prospect. Sci. China Ser. Earth Sci. 2009, 39, 1633–1658. [Google Scholar]
  2. Panakkat, A.; Adeli, H. Recurrent neural network for approximate earthquake time and location prediction using multiple seismicity indicators. Comput.-Aided Civ. Infrastruct. Eng. 2009, 24, 280–292. [Google Scholar] [CrossRef]
  3. Forrest, S.; Perelson, A.S.; Allen, L.; Cherukuri, R. Self-Nonself Discrimination in a Computer; IEEE: Piscataway, NJ, USA, 1994. [Google Scholar]
  4. Fernández-Gómez, M.J.; Asencio-Cortés, G.; Troncoso, A.; Martínez-Álvarez, F. Large earthquake magnitude prediction in chile with imbalanced classifiers and ensemble learning. Appl. Sci. 2017, 7, 625. [Google Scholar] [CrossRef] [Green Version]
  5. Beroza, G.C.; Segou, M.; Mousavi, S.M. Machine learning and earthquake forecasting next steps. Nat. Commun. 2021, 7, 4761. [Google Scholar] [CrossRef] [PubMed]
  6. Jain, R.; Nayyar, A.; Arora, S.; Gupta, A. A comprehensive analysis and prediction of earthquake magnitude based on position and depth parameters using machine and deep learning models. Multimed. Tools Appl. 2021, 80, 28419–28438. [Google Scholar] [CrossRef]
  7. Dehbozorgi, L.; Farokhi, F. Effective feature selection for short-term earthquake prediction using neuro-fuzzy classifier. In Proceedings of the Second IITA International Conference on Geoscience and Remote Sensing, Qingdao, China, 28–31 August 2010. [Google Scholar]
  8. Ikram, A.; Qamar, U. A Rule-Based Expert System for Earthquake Prediction. J. Intell. Inf. Syst. 2014, 43, 205–230. [Google Scholar] [CrossRef]
  9. Pandit, A.; Biswal, K.C. Prediction of earthquake magnitude using adaptive neuro fuzzy inference system. Earth Sci. Inform. 2019, 9, 513–524. [Google Scholar] [CrossRef]
  10. Luis, S. Predict the magnitudes of seismic events using Bayesian methods. Soil Dyn. Earthq. Eng. 2020, 129, 105914. [Google Scholar] [CrossRef]
  11. Panakkat, A.; Adeli, H. Neural network models for earthquake magnitude prediction using multiple seismicity indicators. Int. J. Neural Syst. 2007, 17, 13–33. [Google Scholar] [CrossRef]
  12. Adeli, H.; Panakkat, A. A probabilistic neural network for earthquake magnitude prediction. Neural Netw. 2009, 22, 1018–1024. [Google Scholar] [CrossRef]
  13. Shi, C. Application of neural network to earthquake engineering. Earthq. Eng. Eng. Vib. 1991, 11, 39–46. [Google Scholar]
  14. Asim, K.M.; Idris, A.; Iqbal, T.; Martínez-Álvarez, F. Earthquake prediction model using support vector regressor and hybrid neural networks. PLoS ONE 2018, 13, e0199004. [Google Scholar] [CrossRef] [PubMed]
  15. Morales-Esteban, A.; Martínez-Álvarez, F.; Troncoso, A.; Justo, J.L.; Rubio-Escudero, C. Pattern recognition to forecast seismic time series. Expert Syst. Appl. 2010, 37, 8333–8342. [Google Scholar] [CrossRef]
  16. Asencio-Cortes, G.; Martinez-Alvarez, F.; Morales-Esteban, A.; Reyes, J.; Troncoso, A. Improving earthquake prediction with principal component analysis application to Chile. In International Conference on Hybrid Artificial Intelligence Systems; Springer: Berlin/Heidelberg, Germany, 2015; pp. 393–404. [Google Scholar]
  17. Asim, K.M.; Moustafa, S.S.; Niaz, I.A.; Elawadi, E.A.; Iqbal, T.; Martínez-Álvarez, F. Seismicity analysis and machine learning models for short-term low magnitude seismic activity predictions in Cyprus. Soil Dyn. Earthq. Eng. 2020, 130, 105932. [Google Scholar] [CrossRef]
  18. DeVries, P.M.; Viégas, F.; Wattenberg, M.; Meade, B.J. Deep learning of aftershock patterns following large earthquakes. Nature 2018, 560, 632. [Google Scholar] [CrossRef]
  19. Huang, J.P.; Wang, X.A.; Zhao, Y.; Xin, C.; Xiang, H. Large earthquake magnitude prediction in Taiwan based on deep learning neural network. Neural Netw. World 2018, 28, 149–160. [Google Scholar] [CrossRef]
  20. Wang, Q.; Guo, Y.; Yu, L.; Li, P. Earthquake prediction based on spatio-temporal data mining: An LSTM network approach. IEEE Trans. Emerg. Top. Comput. 2020, 8, 148–158. [Google Scholar] [CrossRef]
  21. Wu, J.; Liang, Y.; Tan, C.; Zhou, W. Method of earthquake prediction based on negative selection. Appl. Res. Comput. 2019, 36, 1097–1100. [Google Scholar]
  22. Gan, Y.; Liang, Y.; Tan, C.Y.; Zhou, W.; Wu, J.J. Earthquake prediction method based on danger theory. Comput. Eng. 2019, 46, 278–283. [Google Scholar]
  23. Zhou, W.; Dong, H.; Liang, Y. The deterministic dendritic cell algorithm with Haskell in earthquake magnitude prediction. Earth Sci. Inform. 2020, 13, 447–457. [Google Scholar] [CrossRef]
  24. Asencio-Cortes, G.; Martinez-Alvarez, F.; Morales-Esteban, A.; Troncoso, A. Medium-large earthquake magnitude prediction in tokyo with artificial neural networks. Neural Comput. Appl. 2017, 28, 1043–1055. [Google Scholar] [CrossRef]
  25. Alizadeh, E.; Meskin, N.; Khorasani, K. A dendritic cell immune system inspired scheme for sensor fault detection and isolation of wind turbines. IEEE Trans. Ind. Inform. 2017, 14, 545–555. [Google Scholar] [CrossRef]
  26. Center CEN. National Earthquake Data Center. 2020. Available online: http://data.earthquake.cn (accessed on 30 December 2020).
Figure 1. The DC-NSA architecture for EQP.
Figure 1. The DC-NSA architecture for EQP.
Electronics 12 00009 g001
Figure 2. Box plot of the eight seismic indicators distribution of each dataset.
Figure 2. Box plot of the eight seismic indicators distribution of each dataset.
Electronics 12 00009 g002
Table 1. Calculation methods of each seismic indicator.
Table 1. Calculation methods of each seismic indicator.
IndicatorCalculation Methods
b i ( n Σ ( M i l o g N i ) Σ M i ( Σ l o g N i ) ) / ( ( Σ M i ) 2 n Σ M i 2 )
η i ( Σ ( l o g 10 N i ( a i b M i ) 2 ) ) / ( n 1 )
where a i = Σ ( l o g 10 N i + b M i ) / n
Δ M i M m a x , o b s e r v e d i M m a x , e x p e c t e d i
T M s i t i t 1
μ M s i Σ ( t i ) / n c
c M s i s t a n d a r d d e v i a t i o n o f t h e o b s e r v e d t i m e s / u M s i
d E i 1 / 2 E i = 10 11.8 + 1.5 M i e r g s
d E i 1 / 2 = Σ E i 1 / 2 / T M s i
M m e a n i Σ M i / n
Table 2. Signal weight values.
Table 2. Signal weight values.
PSDSSS
CSM0.40.20.4
SEMI00−1
MAT0.40.20.4
Table 3. Analyzed seismic areas of China.
Table 3. Analyzed seismic areas of China.
AreasNumber of DataMeanSD
Gansu (DS1)11323.510.50
Qinghai (DS2)20863.670.61
Sichuan (DS3)54333.520.49
Yunnan (DS4)33893.470.49
Table 4. Partial eight indicators and label of Sichuan province after preprocess.
Table 4. Partial eight indicators and label of Sichuan province after preprocess.
Dateb η Δ M T μ c dE 1 / 2 ( × 10 10 ergs ) M mean C
2018100.770.001−0.21267201.440.0153.270
2016070.810.0040.66167140.670.0323.1511
..............................
2011091.020.005−0.2187100.610.0333.290
2001020.750.003−0.0121003.5553.0070
Table 5. Comparison between DC-NSA and other approaches when predicting DS1. The best results are in bold.
Table 5. Comparison between DC-NSA and other approaches when predicting DS1. The best results are in bold.
PPVNPVRnSFARMCCAUCAvgR
DC-NSA0.430.730.710.450.230.140.590.59 0.50
DCA [22]0.430.750.720.530.250.180.600.610.47
NSA [21]0.210.520.210.510.380.020.510.36−0.17
PCA-RF [16]0.550.610.600.210.420.230.520.490.18
BPNN [11]0.640.630.350.150.370.230.610.44−0.02
RNN [2]0.270.750.680.320.240.040.480.510.44
PNN [12]0.320.640.130.630.33−0.020.510.48−0.20
EQP-hDCA [23]0.320.740.700.370.260.060.530.530.44
LSTM [19]0.420.810.690.310.230.120.540.560.46
SVR-HNN [14]0.380.760.670.290.280.090.510.530.39
Table 6. Comparison between DC-NSA and other approaches when predicting DS2. The best results are in bold.
Table 6. Comparison between DC-NSA and other approaches when predicting DS2. The best results are in bold.
PPVNPVRnSFARMCCAUCAvgR
DC-NSA0.841.001.000.960.000.910.970.951.00
DCA [22]0.170.820.500.520.18−0.010.460.500.32
NSA [21]0.000.510.490.000.420.000.450.25−0.42
PCA-RF [16]0.610.560.580.770.410.020.430.630.37
BPNN [11]0.670.960.670.040.040.630.820.580.63
RNN [2]0.860.980.900.910.010.870.900.910.89
PNN [12]0.840.920.810.820.060.810.860.850.75
EQP-hDCA [23]0.731.001.000.930.000.830.970.921.00
LSTM [19]0.780.820.910.870.060.820.850.850.85
SVR-HNN [14]0.690.800.860.790.120.680.660.790.74
Table 7. Comparison between DC-NSA and other approaches when predicting DS3. The best results are in bold.
Table 7. Comparison between DC-NSA and other approaches when predicting DS3. The best results are in bold.
PPVNPVRnSFARMCCAUCAvgR
DC-NSA0.911.001.000.910.000.920.960.931.00
DCA [22]0.330.780.240.150.220.100.620.370.02
NSA [21]0.330.690.730.490.280.180.620.560.45
PCA-RF [16]0.320.730.430.210.320.120.440.420.11
BPNN [11]0.560.900.500.080.100.440.710.510.40
RNN [2]0.450.660.660.550.330.020.520.580.14
PNN [12]0.470.790.640.590.190.380.510.620.45
EQP-hDCA [23]0.460.910.820.630.090.410.730.710.73
LSTM [19]0.650.790.810.530.110.650.710.700.70
SVR-HNN [14]0.590.620.730.490.190.590.630.610.54
Table 8. Comparison between DC-NSA and other approaches when predicting DS4. The best results are in bold.
Table 8. Comparison between DC-NSA and other approaches when predicting DS4. The best results are in bold.
PPVNPVRnSFARMCCAUCAvgR
DC-NSA0.770.940.950.850.010.790.900.850.96
DCA [22]0.330.580.500.590.42−0.090.430.500.08
NSA [21]0.000.320.320.000.550.000.230.16−0.55
PCA-RF [16]0.310.590.530.230.430.020.500.410.10
BPNN [11]1.000.100.150.000.900.120.590.31−0.75
RNN [2]0.630.670.590.600.280.550.570.620.31
PNN [12]0.660.680.530.610.290.460.550.620.24
EQP-hDCA [23]0.670.980.970.770.020.700.870.850.95
LSTM [19]0.350.310.480.420.210.570.460.390.37
SVR-HNN [14]0.370.560.610.590.160.650.560.530.45
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, W.; Lan, W.; Ye, Z.; Ming, Z.; Chen, J.; He, Q. Adaptive Dendritic Cell-Negative Selection Method for Earthquake Prediction. Electronics 2023, 12, 9. https://doi.org/10.3390/electronics12010009

AMA Style

Zhou W, Lan W, Ye Z, Ming Z, Chen J, He Q. Adaptive Dendritic Cell-Negative Selection Method for Earthquake Prediction. Electronics. 2023; 12(1):9. https://doi.org/10.3390/electronics12010009

Chicago/Turabian Style

Zhou, Wen, Wuyang Lan, Zhiwei Ye, Zhe Ming, Jingliang Chen, and Qiyi He. 2023. "Adaptive Dendritic Cell-Negative Selection Method for Earthquake Prediction" Electronics 12, no. 1: 9. https://doi.org/10.3390/electronics12010009

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop