Next Article in Journal
Enhanced Information Graph Recursive Network for Traffic Forecasting
Next Article in Special Issue
Improved BIGRU Model and Its Application in Stock Price Forecasting
Previous Article in Journal
A Projected AR Serious Game for Shoulder Rehabilitation Using Hand-Finger Tracking and Performance Metrics: A Preliminary Study on Healthy Subjects
Previous Article in Special Issue
Improved Design and Application of Security Federation Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Study on Downhole Geomagnetic Suitability Problems Based on Improved Back Propagation Neural Network

1
College of Science, North China University of Science and Technology, Tangshan 063210, China
2
College of Mining Engineering, North China University of Science and Technology, Tangshan 063210, China
3
College of Artificial Intelligence, North China University of Science and Technology, Tangshan 063210, China
4
College of Economics, North China University of Science and Technology, Tangshan 063210, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(11), 2520; https://doi.org/10.3390/electronics12112520
Submission received: 16 April 2023 / Revised: 28 May 2023 / Accepted: 29 May 2023 / Published: 2 June 2023
(This article belongs to the Special Issue Intelligent Analysis and Security Calculation of Multisource Data)

Abstract

:
The analysis of geomagnetic suitability is the basis and premise of geomagnetic matching navigation and positioning. A geomagnetic suitability evaluation model using mixed sampling and an improved back propagation neural network (BPNN) based on the gray wolf optimization (GWO) algorithm by incorporating the dimension learning-based hunting (DLH) search strategy algorithm was proposed in this paper to accurately assess the geomagnetic suitability. Compared with the traditional geomagnetic suitability evaluation model, its generalization ability and accuracy were better improved. Firstly, the key indicators and matching labels used for geomagnetic suitability evaluation were analyzed, and an evaluation system was established. Then, a mixed sampling method based on the synthetic minority over-sampling technique (SMOTE) and Tomek Links was employed to extend the original dataset and construct a new dataset. Next, the dataset was divided into a training set and a test set, according to 7:3. The geomagnetic standard deviation, kurtosis coefficient, skewness coefficient, geomagnetic information entropy, geomagnetic roughness, variance of geomagnetic roughness, and correlation coefficient were used as input indicators and put into the DLH-GWO-BPNN model for model training with matching labels as output. Accuracy, recall, the ROC curve, and the AUC value were taken as evaluation indexes. Finally, PSO (Particle Swarm Optimization)-BPNN, WOA (Whale Optimization Algorithm)-BPNN, GA (Genetic Algorithm)-BPNN, and GWO-BPNN algorithms were selected as compared methods to verify the predictable ability of the DLH-GWO-BPNN. The accuracy ranking of the five models on the test set was as follows: PSO-BPNN (80.95 %) = WOA-BPNN (80.95%) < GA-BPNN (85.71%) = GWO-BPNN (85.71%) < DLH-GWO-BPNN (95.24%). The results indicate that the DLH-GWO-BPNN model can be used as a reliable method for underground geomagnetic suitability research, which can be applied to the research of geomagnetic matching navigation.

1. Introduction

Navigation has a long history of development and plays an indispensable role in human production and life. Traditional navigation systems such as satellite and inertial navigation can no longer meet the requirements of long-endurance and high-precision navigation development in applications [1]. Geomagnetic navigation uses the features of the geomagnetic field for navigation and positioning, which has the advantages of being passive, all-day, and all-weather [2,3,4,5]. Additionally, it has become a new method for combined positioning of indoor and underground engineering [6].
The accuracy of geomagnetic matching navigation positioning is related to a large extent to the suitability of the geomagnetic map [7]. The suitability of the geomagnetic map represents the corresponding relationship between geomagnetic features and geographical location. The suitability directly affects the accuracy of geomagnetic matching navigation and positioning [8,9]. The matching area with good suitability contains more abundant geomagnetic features and information [10,11]. The analysis of geomagnetic suitability is the basis and premise of geomagnetic matching navigation positioning [12]. Therefore, the study of geomagnetic suitability analysis is of great significance.
In the current literature, researchers mainly improve the geomagnetic suitability evaluation method by constructing comprehensive suitability feature parameters or introducing intelligent classification algorithms.
Chen Y.R. et al. [13] used the fractal dimension as the feature parameter of geomagnetic suitability to evaluate the suitability. The outcomes demonstrated that this method had the advantages of a small calculation, good anti-interference, and simple implementation. Wang X.L. et al. [14] compared the selection methods of matching area based on geomagnetic suitability feature parameters and based on the information entropy of geomagnetic suitability feature parameters. The latter was found to be more effective.
Zhao J.H. et al. [15] proposed a matching area division method for underwater geomagnetic navigation based on the geomagnetic symbiosis matrix. This method can respond to the changing features of the geomagnetic field in multiple directions. Zhu Z.L. et al. [16] studied the sensitivity of multi-attribute weight in geomagnetic suitability evaluation using the Weighted Product Model (WPM) method. This study obtained the order of sensitivity of geomagnetic suitability indicators, which had certain reference significance for setting the geomagnetic map suitability feature weights.
Huang Z.X. et al. [17] adopted statistical parameters such as entropy and roughness as geomagnetic feature parameters. The correlation analysis of the feature parameters was carried out based on the matching test, and then the performance analysis of the matching area was realized. Liu Y.X. et al. [18] presented a new approach to selecting geomagnetic matching areas by integrating multiple feature parameters. The proposed approach was proven to be feasible through the simulation test. Li D.W. et al. [19] proposed a method to construct a comprehensive evaluation value based on factor analysis and entropy methods for suitability evaluation. The matching algorithm was used to simulate the experiment. The experiment verified the high agreement between comprehensive evaluation value and matching probability.
Wang Z. et al. [20] defined the concept of credibility and analyzed the credibility between four commonly used geomagnetic feature parameters and matching probability. It was found that a single feature parameter cannot be used as an effective basis for geomagnetic suitability evaluation. Wang P. et al. [21] comprehensively considered four characteristic parameters of a geomagnetic map. The decision method containing maximum deviation and entropy was used to construct the comprehensive evaluation value. The experiment was carried out to prove that the comprehensive evaluation value can be used as a quantitative basis for geomagnetic suitability analysis. Wang L.H. et al. [22] proposed a fuzzy, vague set evaluation method with more information based on the fuzzy decision method. Based on this method, the comprehensive evaluation value was constructed, and the simulation experiment was carried out. It was found that the matching error of the region with a large comprehensive evaluation value was small. Zhu Z.L et al. [23] fused the five indicators and used entropy technology to modify the weight of each indicator obtained based on traditional fuzzy evaluation methods to obtain the comprehensive evaluation value. The traditional Mean Square Differences (MSD) and Mean Absolute Differences (MAD) matching algorithms were used for simulation experiments. It was found that the comprehensive evaluation value and the matching probability were highly consistent, and the comprehensive evaluation value can comprehensively evaluate the geomagnetic suitability.
Zhong Y. et al. [24] used five main feature parameters as fuzzy indicators for weighted analysis to obtain a comprehensive evaluation value. An experimental study was conducted on the basis of the non-tracking Kalman filter with geomagnetic anomalies and the anomaly grid data of some waters in the South China Sea. Experiments showed that the proposed method was reliable. Zhang H. et al. [25] established a comprehensive evaluation model combining Principal Component Analysis (PCA) and Analytic Hierarchy Process (AHP) algorithms, which was used to evaluate the suitability. He Y.J. et al. conducted a comprehensive geomagnetic suitability evaluation based on AHP, the information entropy method [26], and the gray correlation method [27], respectively. The correlative matching algorithm was used to test it, and it was found that the comprehensive evaluation value and the matching probability are highly consistent.
Related experts and scholars on geomagnetic suitability evaluation have done a lot of work and achieved fruitful results by constructing the comprehensive evaluation value. The index system of this research is relatively simple, and the workload is large and cumbersome. There are a few types of research on suitability evaluation using classification algorithms. Additionally, most of them adopt the BPNN algorithm or an optimized BPNN algorithm. Zhang K. et al. [28] suggested automatic recognition and division methods for background field matching/mismatch areas based on BPNN. On the basis of the BPNN algorithm, Wang C.Y. et al. [29] combined PCA and GA to gain improved methods to evaluate geomagnetic suitability and improve classification accuracy. Wang J.H. et al. [30] studied the suitability of downhole geomagnetism by using the BPNN algorithm based on contribution factors. The accuracy of training sets was 95%, and the accuracy of test sets was close to 73%. The intelligent classification algorithm for suitability evaluation can decrease the subjective impact. At the same time, more geomagnetic suitability feature indicators can be involved in the calculation, making the evaluation results more comprehensive and improving the efficiency of geomagnetic suitability evaluation. It will be the focus of future geomagnetic suitability research.
The GWO algorithm is a population intelligence algorithm that has the characteristics of simple operation, few parameters, and easy implementation. It is often used to optimize the BPNN algorithm and is applied in engineering practice. Bao W. et al. [31], based on the battery operation data of the electric vehicle cloud platform with a sampling period of 10 s, used the data to test the unoptimized BPNN, GWO-BPNN, and PSO-BPNN. The experimental results indicate that the GWO-BPNN has high accuracy in predicting the SOC of electric vehicle batteries. Jing W.Q. et al. [32] proposed an improved GWO-BPNN energy consumption prediction model, which was validated using actual operating data of an office building in Xi’an. The prediction results showed that the prediction accuracy of the improved GWO-BPNN model was much higher than that of traditional prediction models. Guo Cui et al. [33] used the gray wolf population algorithm to optimize the BPNN to establish a dynamic model of the accelerometer and simulate the input and output signals. The results show that, compared with the BPNN algorithm, the algorithm has improved its solution accuracy by 43.4% after optimization and improvement. However, due to its tendency to fall into local optima, the GWO algorithm needs to be improved to achieve a balance between local and global search. Li Z. et al. [34] proposed a binary version of the local adversarial learning golden sine gray wolf optimization algorithm and verified its high search accuracy. Pan H. et al. [35] proposed an improved gray wolf optimization algorithm for feature selection of high-dimensional data. Pan T et al. [36] integrated the nonlinear convergence factor and position mutation strategy into GWO to solve the problem of complex and non-linear components and prevent the gray wolf algorithm from falling into a local optimum.
Traditional suitability evaluation methods were mostly applied in the air and underwater. There was little research on downhole suitability evaluation methods. To select more suitable areas for matching, this paper constructs an intelligent model for geomagnetic suitability evaluation based on multi-algorithm coupling. Firstly, a mixed sampling method was constructed by combining SMOTE and Tomek Links to process the original datasets collected from 41 underground engineering projects and obtain a new dataset. Secondly, a hybrid optimization algorithm based on the DLH algorithm and the GWO algorithm is used to optimize the parameters of the BPNN algorithm, and the DLH-GWO-BPNN algorithm is used to construct a geomagnetic suitability model. Then, using accuracy, recall, AUC value, and ROC curve as evaluation indicators, PSO-BPNN, WOA-BPNN, GA-BPNN, and GWO-BPNN algorithms were selected as compared methods to verify the predictable ability of the DLH-GWO-BPNN. The results showed that this model can effectively evaluate geomagnetic suitability and has reference significance for intelligent geomagnetic navigation.
The remainder of this paper is organized as follows: The GWO algorithm and the BPNN algorithm are described in Section 2; the DLH-GWO-BPNN model is shown in Section 3; Section 4 shows the data processing results and the analysis results; and the conclusion is given in Section 5.

2. Related Work

2.1. GWO Algorithm

The gray wolf optimization (GWO) algorithm is an optimization method for swarm intelligence [37]. It simulates the leading status and relatively complete hunting process of gray wolves in their natural environment. Additionally, it is widely used because of its simple structure, few adjustable parameters, and strong convergence performance.
The algorithmic content is as follows:
In GWO algorithm, there are four levels of wolves, as illustrated in Figure 1. α wolf is the head gray wolf. α wolf, as the leader, is in charge of all kinds of activities among the wolves and is the most capable of managing them. β wolf obeys the orders of α, can dominate other gray wolves, and is the best successor of the head gray wolf. δ wolf assists the decision-making of the β wolf and obeys the management of the α wolf and β wolf. The ω wolf is the last level of the social hierarchy, and its presence effectively prevents internal problems such as fighting each other.
The GWO algorithm includes five parts:
(1)
Social level
In GWO algorithm, there are four levels of gray wolf population: α, β, δ and ω. The rank of society is α, β, δ, and ω. The α, β, and δ are also the three gray wolves with the best fitness in each generation’s population, which are constantly updated in the iterative process of the GWO algorithm.
(2)
Surrounding prey
α, β, and δ directly capture prey. ω wolf surrounds prey. Its surrounding process is described by Formulas (1) and (2).
Q = C × Y p ( t ) Y ( t )
Y ( t + 1 ) = Y p ( t ) A × Q
Formula (1) is the distance between any gray wolf in the gray wolf population and prey. Formula (2) is the updated formula for the offspring of the gray wolves. Y(t) and Y(t + 1) represent the positions of gray wolves after the tth iteration and the (t + 1)th iteration, respectively. Yp(t) is the position of prey. A and C are the coefficients, which are obtained according to Formulas (3) and (4):
A = 2 a × n 2 a
C = 2 n 1
a is a control factor, and it is linear. n1 is a random number between 0 and 1. The value range of n2 is within [0, 1].
(3)
Hunting
α, β, and δ are the best three wolves. They find the location of prey, notify ω, and round up the prey. The α, β, and δ determine the location of prey. ω, according to the positions of α, β, and δ, constantly adjusts its position. The mathematical principle is described by the following formula:
Q α = C 1 × Y α ( t ) Y ( t )
Q β = C 2 × Y β ( t ) Y ( t )
Q δ = C 3 × Y δ ( t ) Y ( t )
Qα, Qβ, and Qδ are the distances between α, β, and δ wolf and other individuals, respectively. Yα(t), Yβ(t), and Yδ(t) represent the positions of α, β, and δ after the tth iteration, respectively. Ci is an adaptive vector.
Y 1 = Y α A 1 × Q α
Y 2 = Y β A 2 × Q β
Y 3 = Y δ A 3 × Q δ
Y t + 1 = Y 1 + Y 2 + Y 3 3
Y1, Y2, and Y3 represent the direction and distance of the ω wolf to the α wolf, β wolf, and δ wolf, respectively. Formula (11) is the final position of the wolf pack. Ai is a random vector.
According to the above formulas, gray wolves track and round up prey, as shown in Figure 2:
(4)
Attacking prey
When the prey is still in place, the whole wolf pack attacks. The process of attacking prey in GWO is simulated. It can be seen from Formula (3) that as a decreases, A also decreases continuously. In the iteration process, a decreases from 2 to 0, which is a linear change process, and A also changes in the corresponding interval. A controls the expansion and contraction of the encirclement circle of the gray wolf pack. When |A| > 1, the whole wolf pack will move far away from the prey in order to find optimal prey; that is global search. When |A| < 1, the whole wolf pack will move to the location of the prey and launch an attack; that is local search. The schematic diagram of this process is shown in Figure 3.
(5)
Searching prey
When |A| > 1, they will go away from the prey, then globally search for other better prey. Because of the randomness of the parameter C in finding prey, the algorithm has a stochastic search behavior. In the algorithm design, it has a certain avoidance effect on falling into the local optimum. It also acts as an obstacle in nature that prevents the wolf population from capturing prey easily.

2.2. BPNN Algorithm

BPNN algorithm has strong fault tolerance and generalization ability, which are applied in many fields [38]. The structure of BPNN has three sections, the configuration of which is illustrated schematically in Figure 4. The neurons in the input layer can obtain the dataset. The hidden layer is located in the middle of the input and output layers. The dataset of the input layer is transformed and processed by the hidden layer, and then sent to the output layer to obtain the final result.
The mathematical principles of BPNN are divided into the following two parts:
(1)
Forward propagation
Forward propagation means that the neural network calculates and stores the intermediate variables in sequence, from input layer to hidden layer and then to output layer. After the parameters of BPNN are initialized, the training data are input from the input layer and then transmitted to the hidden layer, and the input of the neurons in the hidden layer is obtained.
Z j = i = 1 n v i j x j j = 1 , 2 , , q
Zj represents the input value of the jth neuron of the hidden layer. Xi represents the input value of the ith neuron of the input layer, namely the sample data. vij represents the connection weight.
The output values of neurons in the hidden layer are calculated according to the thresholds and the input values of the neurons in the hidden layer. The calculation formula for the output value is:
M j = f 1 ( Z j a j ) j = 1 , 2 , , q
f1 represents the transfer function. aj is the threshold of the jth neuron in the hidden layer. Mj is the output value of the jth neuron in the hidden layer.
According to the output value Mj of the hidden layer and the given weights, calculate the results and transfer them to the output layer as input values. The input of the output layer is obtained.
O k = j = 1 q w j k M j k = 1 , 2 , , m
Ok represents the input value of the kth neuron in the output layer. wjk represents the weight.
According to the thresholds of neurons in the output layer and the input value of the output layer, the output value of the neurons in the output layer is calculated and obtained. The calculation formula for the output value is:
N k = f 2 ( O k b k ) k = 1 , 2 , , m
Nk is the output value of the kth neuron in the output layer. f2 is the transfer function. bk is the threshold of the kth neuron in the output layer.
(2)
Back propagation
Back propagation refers to the calculation method of BPNN parameter gradient. According to the chain rule of calculus, the gradient of loss function for each parameter in the BPNN algorithm is calculated in sequence from the output to the input, and the parameters are updated with the optimization method to reduce the loss function.
According to Nk and Yk of each neuron, the error of BPNN is obtained.
E = 1 2 k = 1 m ( Y k N k )
E is the error in BPNN. Yk is the expected output value. Nk is the actual output value.
When the error is not within the acceptable range, vij, wjk, aj, and bk are adjusted according to the gradient descent theory and chain rule. The correction values of vij, wjk, aj, and bk are obtained.
Δ v ij = η E v i j = η ( E M j ) ( M j v i j ) = η ( E M j ) ( M j Z j ) ( Z j v i j )
Δ v j k = η E w j k = η ( E N k ) ( N k w j k ) = η ( E N k ) ( N k O k ) ( O k w j k )
Δ a j = η E a j = η ( E M j ) ( M j a j )
Δ b k = η E b k = η ( E N k ) ( N k b k )
η indicates learning rate. The value range of η is within [0, 1].
The revised weights and thresholds are:
v ij = v ij + Δ v ij w jk = w jk + Δ w jk a j = a j + Δ a j b k = b k + Δ b k
After the weights and thresholds are updated, forward propagation is carried out to make the actual output approximate the theoretical output to the greatest extent.

2.3. The SMOTE and Tomek Links Method

(1)
SMOTE
Synthetic Minority Oversampling Technique (SMOTE) is a classic oversampling method to deal with data imbalance. A synthetic sample is generated by taking the corresponding line segment of two minority samples as the endpoints to increase the number of minority samples and achieve the purpose of oversampling minority samples. The principle of SMOTE in making the new sample is shown in Formula (22).
x n e w = x + r a n d ( 0 , 1 ) × x x
xnew represents a new sample; x is a minority sample; x means the nearest neighbor sample.
(2)
Tomek Links
After oversampling, if the last two samples belong to two categories, the two samples are combined into a Tomek Links pair. One of the samples is noise data, and this sample is deleted. Tomek Links can effectively delete the overlapping data, which enables the classifier to classify better.

2.4. Evaluation Indicators of Model Performance

(1)
Accuracy
For a given set of samples, the calculation formula for accuracy is defined as follows:
A ccuracy = T P + T N T P + T N + F P + F N
TP represents positive samples predicted by the model as positive classes; TN represents negative samples predicted by the model as negative classes; FP represents negative samples predicted by the model as positive; FN represents positive samples predicted by the model as negative classes.
(2)
Recall
The calculation formula for recall is shown in Equation (24).
Recall = T P T P + F N
(3)
ROC curve and AUC
ROC (Regional Operating Characteristic) curve and AUC (Area Under Curve) are indicators for the comprehensive evaluation of binary classification problem. ROC curve provides a new graphical measurement of algorithm performance. The curve, which is close to the upper left corner, indicates the corresponding algorithm has good performance. AUC is the area under the ROC curve, and its value is within [0, 1]. The value that is close to 1 indicates an excellent result. This value can quantify the performance of the classification algorithm. In this paper, the idea of the ROC curve applied to multi-class classification is to plot the average ROC curve. Convert the sample label to a binary-like form, and then calculate the probability value of each sample under each label, and the average ROC curve is obtained.

3. DLH-GWO-BPNN Model

The GWO algorithm based on DLH is applied to optimize BPNN and construct the DLH-GWO-BPNN model.

3.1. Improved GWO Algorithm Based on DLH

The candidate positions in the GWO algorithm are mainly determined by α, β, and δ, which easily fall into the local optimal solution. To keep the balance between local and global search, the DLH algorithm is introduced to optimize the GWO algorithm. The inspiration for DLH comes from the hunting behavior of individuals in nature [39]. In the GWO algorithm, the concept of neighborhood is introduced to provide a new choice of candidate location for every gray wolf. The improved GWO algorithm based on DLH consists of three stages: the initialization phase, the movement phase, and the selection and update phases.
(1)
Initialization
According to Formula (25), determine the scope of the search space, in which N gray wolves are randomly initialized.
Y km = L k + r a n d m [ 0 , 1 ] × u k l m   k [ 1 , N ] , m [ 1 , D ]
Yk(t) = {yk1, yk2, …ykD} denotes the position of the kth gray wolf after the tth iteration. D is the dimension of the problem. The fitness function is f(x). Calculate Yk(t) fitness according to f(x).
(2)
Movement
Individual hunting is not only an important behavior in the whole gray wolf population but also the inspiration for improving the GWO algorithm. In the DLH algorithm, every gray wolf is considered by their neighbors to be a candidate position after the tth iteration. The conventional GWO algorithm and the DLH algorithm generate new candidate positions as follows:
GWO: On the basis of the positions of α wolf, β wolf, and δ wolf and the calculated coefficients a, A, and C, determine the encirclement of the wolves. Then calculate the new candidate position Yk-GWO(t + 1) after the tth iteration of the kth gray wolf according to Formula (11).
DLH: Each gray wolf learns from its neighbors and other random individuals, resulting in the new candidate position Yk-DLH(t + 1) in the following steps.
Step 1: Calculate the euclidean distance between the current position and the candidate position Yk-GWO(t + 1) of each gray wolf of the GWO algorithm. According to this distance, calculate the search radius Rk(t) of every gray wolf.
R k t = [ Y k ( t ) Y k - G W O ( t + 1 ) ] 2
Step 2: Determine the neighborhood of Yk(t) with the following expression.
N k ( t ) = Y m ( t ) Q k ( Y k ( t ) , Y m ( t ) ) R k ( t ) , Y m ( t ) P o p
Qk denotes the Euclidean distance between Yk(t) and Ym(t). Pop represents a matrix that stores wolves. It has N rows and D columns.
Step 3: Select randomly the neighborhood Yn,d(t) in Nk(t). Select randomly Yr,d(t) from Pop. The new candidate position Yk-DLH(t + 1) is calculated according to these two quantities.
Y k - D L H , d ( t + 1 ) = Y k , d ( t ) + r a n d × ( Y n , d ( t ) Y r , d ( t ) )
(3)
Selection and updating
The excellent candidate position is selected by calculating and comparing the fitness of the two positions with the following mathematical expression.
Y k ( t + 1 ) = Y k - G W O ( t + 1 ) i f f ( Y k - G W O ) < f ( Y k - D L H ) Y k - D L H ( t + 1 ) O t h e r w i s e
Based on this principle, update the position of each gray wolf until it reaches the largest number of iterations.

3.2. DLH-GWO-BPNN Model

In the evaluation of geomagnetic suitability, in order to improve the accuracy of the traditional BPNN model, the collected test data are mixed-sampled. After data preparation, the DLH-GWO algorithm is used to optimize the traditional BPNN model and build the DLH-GWO-BPNN model.
The specific steps are as follows:
Step 1: Construction of a mixed sampling method based on the SMOTE and Tomek Links algorithms. The SMOTE algorithm is used to synthesize the unmatched samples, weakly matched samples, and matched samples in the original dataset. Use the Tomek Links algorithm for data cleaning. Then normalize the sampled dataset.
Step 2: Set the maximum number of iterations of the DLH-GWO algorithm. Determine the topology of BPNN.
Step 3: The initial dimension represents the initial weight and threshold of BPNN. Put it into BPNN for training. Calculate the fitness of each gray wolf in the wolf pack. Select the three gray wolves with the lowest fitness value. Ranked from small to large according to fitness, namely α, β, and δ.
Step 4: Update the position of each gray wolf according to Formulas (8)–(11), namely Yk-GWO(t + 1). Then construct a new neural network for training. Recalculate the fitness of each gray wolf. Select the three gray wolves with the lowest fitness value. Ranked from small to large according to fitness, namely α, β, and δ. Update the position of each gray wolf again, namely Yk-GWO(t + 1).
Step 5: Calculate the search radius of each gray wolf according to Formula (26). Then determine the neighborhood of each gray wolf according to Formula (27). Each time a gray wolf learns from its neighbors and other random individuals, a new candidate position is generated, namely Yk-DLH(t + 1).
Step 6: Calculate the fitness of Yk-GWO(t + 1) and Yk-DLH(t + 1). Compare and select excellent candidates for the position.
Step 7: Judge whether the maximum number of iterations of DLH-GWO has been reached. If not, return to Step 3. If so, record α corresponding initial weights and thresholds.
Step 8: Obtain the value range of the number of hidden layer nodes based on empirical formulas. Input the optimal initial weight and threshold into the BPNN for training. Set different numbers of nodes and activation functions for experiments. Carry out a comparative analysis and select the node number and activation function with the smallest error.
Step 9: Take the accuracy, recall, ROC curve, and AUC value as evaluation indicators and put the test set into the trained BPNN for validation analysis.
Traditional geomagnetic suitability evaluation models are mostly based on multiple geomagnetic features to build a comprehensive evaluation value. Then, through simulation experiments, verify the high consistency of the comprehensive evaluation value and the matching probability. Due to the numerous geomagnetic features, the selection of geomagnetic features for the construction of comprehensive evaluation values has greater subjectivity. In order to improve this situation, the data from the matching region is divided into four types of samples with different matching degrees according to the matching probability. The BPNN algorithm is introduced, and the geomagnetic features are taken as input and the matching labels are taken as output to evaluate suitability.
Because the BPNN algorithm is easy to overfit and excessively depends on the initial weight and threshold, the GWO algorithm is introduced to iteratively optimize the initial weight and threshold of the BPNN. The DLH algorithm is introduced to avoid the GWO algorithm falling into a local optimal solution. In addition, due to the imbalance of the proportion of each category of sample data selected, before training, a mixed sampling method is built based on the SMOTE and Tomek Links algorithms for data balance processing. Figure 5 describes the overall concept of the model.

4. Experimental Simulation

4.1. Selection of Indicators

According to the existing research, there are various feature parameters to characterize geomagnetic maps, such as geomagnetic roughness, geomagnetic standard deviation, geomagnetic information entropy, etc. The features of a geomagnetic map are characterized from different angles. Due to the incompleteness of a single feature to characterize geomagnetic suitability, seven indicators are selected from three aspects in combination with the distribution characteristics of the downhole geomagnetic field: macroscopic features, microscopic features, and self-similar features, and the geomagnetic suitability is comprehensively measured. The feature parameters are defined as follows:
The span of latitude and longitude of a candidate geomagnetic field area is set as the B × L grid, and g(k, m) is the geomagnetic strength value at the grid point coordinate (k, m).

4.1.1. Macroscopic Features

(1)
Geomagnetic standard deviation σ
σ is the reflection of the degree of dispersion of geomagnetic data and the general changes of the geomagnetic area in the candidate region. The large σ corresponds to the area with the apparent variation of geomagnetic features, and it is beneficial for the analysis of geomagnetic suitability. It is defined as follows:
σ = 1 B L 1 k = 1 B m = 1 L g k , m g ¯ 2
g ¯ is the average geomagnetic field value, which is defined as follows:
g ¯ = 1 B L k = 1 B m = 1 L g ( k , m )
(2)
Kurtosis coefficient C1
C1 stands for the concentration degree of geomagnetic field data in the candidate region. The larger the C1, the more data are concentrated near the average geomagnetic field value, which is not conducive to matching. It is defined as:
C 1 = 1 B L k = 1 B m = 1 L g k , m g ¯ σ 4 4 3
(3)
Skewness coefficient C2
C2 represents the symmetry of the geomagnetic map and is proportional to the suitability, which is defined as:
C 2 = 1 B L k = 1 B m = 1 L g k , m g ¯ σ 3 3 3

4.1.2. Microscopic Features

(1)
Geomagnetic information entropy H
H reflects the fluctuation features of the geomagnetic field data and the richness of geomagnetic information in the candidate region. The lower the H, the better the matching effect. It is defined as:
H = k = 1 B m = 1 L p ( k , m ) l b p ( k , m ) p ( k , m ) = g ( k , m ) k = 1 B m = 1 L g ( k , m )
(2)
Geomagnetic roughness r
r reflects the local fluctuation state and average smoothness of the geomagnetic field in a candidate region. The greater the geomagnetic roughness, the more abundant the geomagnetic field information in the candidate region. It is defined as:
r = r x + r y 2 r x = 1 B ( L 1 ) k = 1 B m = 1 L 1 g ( k , m ) g ( k , m + 1 ) 2 r y = 1 L ( B 1 ) k = 1 B + 1 m = 1 L g ( k , m ) g ( k , m + 1 ) 2
(3)
Variance of geomagnetic roughness R
R indicates the local fluctuation of geomagnetic data in the candidate region. The more significant the R, the more conducive the analysis of geomagnetic suitability is. It is defined as:
R = r σ

4.1.3. Self-Similar Features

(1)
Correlation coefficient p
The p reflects the independence of geomagnetic field values in the candidate region. A small correlation coefficient corresponds to good matching performance. It is defined as:
p = p x + p y p x = 1 B ( L 1 ) σ 2 k = 1 B m = 1 L 1 g ( k , m ) g ¯ g ( k , m + 1 ) g ¯ p y = 1 L ( B 1 ) σ 2 k = 1 B 1 m = 1 L g ( k , m ) g ¯ g ( k + 1 , m ) g ¯

4.2. Matching Labels

Calculate the matching probability of the selected sample area according to the matching algorithm. According to the matching probability, samples are divided into four kinds of matching labels: 0 (mismatch), 1 (weak match), 2 (match), and 3 (strong match), as shown in Table 1 [30].

4.3. Data Source and Data Processing

The model is trained with geomagnetic feature data and matching labels. The data are the geomagnetic features dataset collected by 41 underground projects of about 3 m, which is derived from the relevant data published in the related literature [30].
The data are processed as follows:
(1)
Mixed sampling
The original data contains five unmatched samples, six weakly matched samples, ten matched samples, and twenty strongly matched samples. This dataset is unbalanced and has a small amount of data. To improve the accuracy of the model, balance the data. SMOTE is used for oversampling the other three types of small samples except for strongly matched samples, and then the Tomek Links algorithm is used to delete overlapping samples between classes.
Finally, 68 samples were obtained, and Table 2 shows part of the samples.
(2)
Normalization
The dataset is normalized to eliminate the impact of dimension.
y = ( M a x y M i n y ) ( x M i n x ) M a x x M i n x + M i n y

4.4. Parameter Selection of BPNN

(1)
Selection of neuron numbers
The number range of neurons in the hidden layer of BPNN is determined according to the empirical Formula (39).
m = n + l + a
m, n, and l represent the number of neurons in the input, hidden, and output layers, respectively. a is an integer within [1,10].
The value within the range is selected as the number of neurons in the hidden layer, respectively. The average accuracy of the training set and test set is obtained through 70 tests to choose the optimal number.
(2)
Selection of the activation function of BPNN
Activation functions are set in the input-hidden layer and hidden-output layer, respectively. The softmax function is suitable for multi-classification neural network output, so the softmax function is selected as the activation function in the hidden-output layer. Carry out ten experiments and take the tansig, softmax, purelin, and radbas functions as activation functions in the input-hidden layer, respectively, then calculate and compare the average accuracy of the training and test sets under the training of the BPNN algorithm.
  • tangsig function
tan s i g ( x ) = 2 1 + e 2 x 1
2.
softmax function
σ ( X i ) = exp ( x i ) j = 1 n exp ( x j )
3.
purelin function
y = x
4.
radbas function
y = exp ( x 2 )

4.5. Experimental Comparison

To verify the predictability of the proposed algorithm, comparative experiments were conducted. The BPNN is optimized by using five different optimization algorithms, namely WOA, GA, PSO, GWO, and DLH-GWO, and the model is constructed. Train the model with geomagnetic features as input and matching labels as output. By adjusting the main hyperparameter of each model, the optimal parameters of each model are obtained, respectively, and then the optimal model is constructed. Place the test set into the optimal model obtained, with geomagnetic features as input and matching labels as output. Evaluate the effect of each model using accuracy, recall, AUC value, and ROC curve as indicators.

5. Results and Analysis

5.1. Selection of Neuron Numbers

According to the empirical Formula (39), the number range of neurons in the hidden layer is within [4,14]. Different numbers of neurons are set, respectively, and experiments are conducted to obtain the corresponding average accuracy. The results are shown in Figure 6.
According to Figure 6, when the number of neurons is 10, the average accuracy in the test set reaches its highest, at 0.9299. Meanwhile, the average accuracy of the training set at this point is relatively high, at 0.9264. When the number of neurons is 12 or 14, the average accuracy in the training set is higher than that when the number of neurons is 10, but the average accuracy of the test set at these two points is lower in comparison. The average accuracy in the test set no longer increases more than that of the number of neurons, which is 10. Thus, the optimal number of neurons is 10.

5.2. Selection of Activation Function

The softmax function is suitable for multi-classification neural network output, so the softmax function is set as the activation function between the hidden layer and the output layer. Take tansig, softmax, purelin, and radbas functions as activation functions between the input layer and the hidden layer, respectively. Carry out 10 experiments and calculate and compare the average accuracy of the DLH-GWO-BPNN model in the training set and test set. The results are shown in Table 3.
Using tansig, softmax, purelin, and radbas functions between the input layer and the hidden layer, respectively, the accuracy of the training set and the test set is different. When the purelin function or softmax function is used, the average accuracy of the test set and training set is below 90%. The softmax function has a good advantage for the output of the multi-classification neural network, but it is not ideal when used as the activation function in the input-hidden layer. When the radbas function is used, the average accuracy of the training set reaches 93.83%, but the average accuracy of the test set is relatively low. When the tansig function is used, the average accuracy of the test set and training set is above 90%, and identification ability is the best in comparison.
Therefore, the tansig function is used in the input-hidden layer, and the softmax function is used in the hidden-output layer.

5.3. Parameter Selection

Through training, the optimal main parameters of five models, namely WOA-BPNN, GA-BPNN, PSO-BPNN, GWO-BPNN, and DLH-GWO-BPNN, were obtained. The results are shown in Table 4:

5.4. Validation Analysis of the Suitability Evaluation Model

(1)
Comparing the predictive ability of models with their accuracy
In order to verify the predictable ability of the DLH-GWO-BPNN model, five geomagnetic suitability evaluation models were established, namely PSO-BPNN, GA-BPNN, WOA-BPNN, GWO-BPNN, and DLH-GWO-BPNN, for comparative analysis. Using accuracy as an indicator, the classification effects of five models are as follows:
From Table 5, it can be seen that the WOA-BPNN model performs the worst on both the training and testing sets. There is a significant difference in accuracy between the PSO-BPNN model training set and the test set. The accuracy of the GA-BPNN model training set and test set is 87.23% and 85.71%, respectively, with a small difference and good generalization ability. The accuracy of the GWO-BPNN model and the DLH-GWO-BPNN model training set is the same, at 91.49%, but the accuracy of the DLH-GWO-BPNN model testing set is much higher than that of the GWO-BPNN model. At the same time, the accuracy of the DLH-GWO-BPNN model training set and test set is not significantly different and is higher than that of the GA-BPNN model. Therefore, the DLH-GWO-BPNN model has the best classification performance.
Visualize the accuracy of the training and testing sets, as shown in Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11:
(2)
Comparing the predictive ability of models from the recall
Using the recall as an indicator to evaluate the classification performance of PSO-BPNN, GA-BPNN, WOA-BPNN, GWO-BPNN, and DLH-GWO-BPNN models, the results are as follows:
From Table 6, it can be seen that the WOA-BPNN model has a poor recall of 0.64 for label 1 in the training set, indicating a poor training effect. However, its recall under label 1 in the test set is 100%. The model is unstable. The WOA-BPNN model has a recall of 20% in the test set and 60% in the training set under label 3, indicating poor model performance. The WOA-BPNN model performs well under the other two types of labels. The recall of the PSO-BPNN model in the training and testing sets under label 3 is 70% and 20%, respectively, indicating poor performance of the model. However, the model has a higher recall under other labels in both the test set and the training set. The recall of the GA-BPNN model in the training and testing sets under label 3 is 60% and 40%, respectively, indicating poor performance of the model. However, the model has a higher recall under other labels in both the test set and the training set. The GWO-BPNN model has a recall of 40% in the test set under label 3, which results in poor performance. The accuracy of the DLH-GWO-BPNN model in both the training and testing sets under four labels is within an acceptable range. The model has the highest recall under lable 0, with a recall of 100% for the test set and training set, respectively. The recall of the test set under label 1 is 75%, and the recall of the training set is 100%, which is acceptable. Overall, the DLH-GWO-BPNN model has the best performance.
(3)
Comparing the predictive ability of models using the AUC and ROC curve
The method of calculating the AUC value of multi-classification is used to calculate the AUC values of five models, namely PSO-BPNN, GA-BPNN, WOA-BPNN, GWO-BPNN, and DLH-GWO-BPNN models. The ROC curves of each model are drawn, as shown in Figure 12 and Figure 13:
From the figure, it can be seen that the AUC values of the WOA-BPNN model in the training and test sets are 0.90 and 0.91, respectively, indicating good model performance. The AUC values of the GA-BPNN model in the training and test sets are 0.95 and 0.96, respectively, indicating good generalization ability of the model. The AUC values of the PSO-BPNN model in the training and test sets are 0.96 and 0.97, respectively. The AUC values of the GWO-BPNN model in the two datasets are 0.94 and 0.93, respectively. The DLH-GWO-BPNN model performed well in the training set with an AUC value of 0.96, and the test set showed the best results with an AUC of 0.99, indicating that the model has excellent generalization ability.
In summary, the DLH-GWO-BPNN model has shown good classification performance on all four indicators.

6. Conclusions and Outlooks

In this paper, an evaluation model based on the data reconstruction model and the DLH-GWO-BPNN classification algorithm was proposed for downhole geomagnetic suitability analysis. The PSO-BPNN, GA-BPNN, WOA-BPNN, and GWO-BPNN algorithms were used to compare and analyze the proposed model, and the conclusions are as follows: A data reconstruction model based on SMOTE and Tomek Links was employed to extend the original data set, remove the sample redundancy, which can also solve the problem of data imbalance and small amounts of data, and provide better data set support for model training. The technology of DLH improved the GWO algorithm to prevent it from falling into a local optimum. The BPNN algorithm was optimized by the improved GWO to enhance its performance. The accuracy, recall, ROC curve, and AUC value were selected as evaluation indicators for model training and testing. The accuracy of the training set and the test set of the model are 91.49 % and 95.24 %, respectively. The recall of the model in the training set is relatively high. The recall labels 0, 2, and 3 in the test set are 1, and the recall of label 1 is 0.75. In addition, the AUC values of the model training set and test set are 0.96 and 0.99, respectively. The comparison of PSO-BPNN, GA-BPNN, WOA-BPNN, GWO-BPNN, and DLH-GWO-BPNN models verifies the good classification effect of the DLH-GWO-BPNN model under different evaluation indicators. The DLH-GWO-BPNN model can efficiently and accurately solve the problem of downhole geomagnetic suitability evaluation and provide a theoretical basis for underground intelligent geomagnetic navigation. Meanwhile, future research will continue to improve sampling methods and algorithm optimization methods, as well as consider deep learning structures to improve the model.

Author Contributions

Writing—review & editing, X.Z. and S.R.; Funding acquisition, X.Z. and L.G.; Writing—original draft, J.L.; Methodology, J.L.; Software, H.M.; Validation, H.M., S.R. and L.G.; Project administration, L.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Postgraduate Innovation Project of North China University of Science and Technology: CXZZBS2021102, Coal spontaneous combustion warning and risk assessment modeling analysis research: 22130209H and the Innovation and Entrepreneurship Project of North China University of Science and Technology: X2022358.

Data Availability Statement

All data generated or analyzed during this study are included in this published article.

Acknowledgments

The authors are thankful to the editor and anonymous referees for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, N.B.; Wang, Y.B.; Wang, Q. Review of research progress in geomagnetic navigation technology. J. Navig. Position 2018, 2, 15–19. [Google Scholar]
  2. Wei, B.W.; Lv, W.H.; Fan, X.J.; Zhu, Y.K.; Guo, Y.J. Current status and outlook of AUV navigation technology development. J. Unmanned. Undersea Syst. 2019, 1, 1–9. [Google Scholar]
  3. Wang, Y.; Fan, R.S.; He, Y.J. Application of improved particle swarm optimization algorithm in geomagnetic matching. Sci. Surv. Mapp. 2021, 5, 51–57+142. [Google Scholar]
  4. Gao, D.; Zhu, M.H.; Han, P. A geomagnetic/inertial depth fusion navigation method. J. Chin. Inert. Technol. 2022, 4, 437–444. [Google Scholar]
  5. Lu, Y.; Wei, D.Y.; Ji, X.C.; Yuan, H. A review of geomagnetic positioning methods. Navig. Position Time 2022, 2, 118–130. [Google Scholar]
  6. Zhang, B.; Wang, J.H.; Guo, Y.F.; Wu, B. Determination of initial weight factors for the BP evaluation method of geomagnetic suitability. Chin. J. Sens. Actuators 2019, 9, 1339–1345. [Google Scholar]
  7. Wang, X.L. Research on several key technologies in geomagnetic matching navigation. Surv. Mapp. Eng. 2011, 1, 1–5. [Google Scholar]
  8. Wang, P.; Hu, X.; Wu, M. A hierarchical decision-making scheme for directional matching suitability analysis in geomagnetic aided navigation. J. Aerosp. Eng. 2014, 10, 1815–1830. [Google Scholar] [CrossRef]
  9. Zhu, C.L.; Pan, M.C.; Zhang, Q.; Liu, Z.Y.; Chen, Z.; Liu, W. Research on the suitability of geomagnetic vector map based on machine learning. Trans. Microsyst. Technol. 2022, 8, 21–24+28. [Google Scholar]
  10. Wang, L.; Yu, L.; Qiao, N.; Sun, D. Analysis and simulation of geomagnetic map suitability based on vague set. J. Navig. 2016, 5, 1114–1124. [Google Scholar] [CrossRef] [Green Version]
  11. Xiao, J.; Qi, X.H.; Duan, X.S.; Wang, J.C. Analysis of geomagnetic navigation direction suitability based on depth convolution neural network. Chin. J. Eng. 2017, 10, 1584–1590. [Google Scholar]
  12. Guo, Y.F.; Wang, J.H.; Li, M.D.; Zhang, B.; Zhang, H.J. Research and analysis of regional suitability of downhole geomagnetic positioning. Progr. Geophys. 2020, 2, 406–414. [Google Scholar]
  13. Chen, Y.R.; Yuan, J.P. Study on the suitability of geomagnetic map based on fractal dimension. Flight Dyn. 2009, 6, 76–79. [Google Scholar]
  14. Wang, X.L.; Su, M.D.; Ding, S.; Shen, J.M. Selection of matching areas in geomagnetic navigation. J. Geodes. Geodyn. 2011, 6, 79–83+88. [Google Scholar]
  15. Zhao, J.H.; Wang, S.P.; Wang, A.X. Selection of underwater geomagnetic navigation matching area based on geomagnetic symbiosis matrix. Geomat. Inform. Sci. Wuhan Univ. 2011, 4, 446–449. [Google Scholar]
  16. Zhu, Z.L.; Li, J. WPM method to analyze the sensitivity of geomagnetic map indicator weights. Com. Eng. Appl. 2017, 13, 60–65. [Google Scholar]
  17. Huang, Z.X.; Luo, X.W.; Guan, Y.L.; Xu, X.B. Analysis of regional suitability of marine geomagnetic navigation. Jiangxi Sci. 2013, 1, 35–38. [Google Scholar]
  18. Liu, Y.X.; Zhou, J.; Ge, Z.L. A selection method of geomagnetic matching region based on projection pursuit. J. Astronuat. 2010, 12, 2677–2682. [Google Scholar]
  19. Li, D.W.; Fan, R.S.; Wang, C.B.; He, Y.J.; Ren, W. Application of factor analysis method in suitability analysis of geomagnetic map. J. Navig. Position 2022, 1, 121–129. [Google Scholar]
  20. Wang, Z.; Wang, S.C.; Zhang, J.S.; Qiao, Y.K.; Chen, L.H. An suitability evaluation method for geomagnetic matching guidance based on AHP. J. Astronaut. 2009, 5, 1871–1878. [Google Scholar]
  21. Wang, P.; Wu, M.P.; Ruan, Q.; Yuan, H.P. Application of multi-attribute decision method in geomagnetic map suitability analysis. Ordnance Ind. Autom. 2011, 8, 65–68. [Google Scholar]
  22. Wang, L.H.; Qiao, N.; Yu, L. A fuzzy inference method for matching area selection in underwater terrain navigation. J. Xidian Univ. 2017, 1, 140–145. [Google Scholar]
  23. Zhu, Z.L.; Yang, G.L.; Shan, Y.D.; Yang, S.J.; Wang, Y.Y. A comprehensive evaluation method for the suitability analysis of geomagnetic map. J. Chin. Inert. Tech. 2013, 3, 375–380. [Google Scholar]
  24. Zhong, Y.; Chai, H.Z.; Liu, F.; Wang, X.; Du, Z.Q. Suitability analysis of geomagnetic map based on fuzzy decision theory. J. Wuhan Univ. 2021, 1, 118–124. [Google Scholar]
  25. Zhang, H.; Yang, L.; Li, M. A novel comprehensive model of suitability analysis for matching area in underwater geomagnetic aided inertial navigation. Math. Probl. Eng. 2019, 2019, 1–11. [Google Scholar] [CrossRef]
  26. He, Y.J.; Fan, R.S.; Wang, Y. Research on geomagnetic suitability analysis method. Sci. Surv. Mapp. 2021, 9, 7–13+48. [Google Scholar]
  27. He, Y.J.; Fan, R.S.; Wang, Y. Application of Grey Relational Analysis in geomagnetic matching navigation. J. Navig. Position 2021, 3, 65–72. [Google Scholar]
  28. Zhang, K.; Zhao, J.H.; Shi, C.; Zhang, H.T. Research on BP neural network for division of underwater terrain matching area. J. Wuhan Univ. 2013, 1, 56–59. [Google Scholar]
  29. Wang, C.Y. A combination of PCA and GA-BP for selection of matching area in geomagnetic navigation. Electron. Opt. Control 2018, 6, 110–114. [Google Scholar]
  30. Wang, J.H.; Zhang, B.; Wu, B.; Guo, Y.F. Study on geomagnetic suitability of BP neural Network based on contribution factor. J. Hefei Univ. Tech. 2020, 12, 1668–1675. [Google Scholar]
  31. Bao, W.; Ren, C. Research on battery SOC prediction method based on GWO-BP neural network. Com. Appl. Softw. 2022, 9, 65–71. [Google Scholar]
  32. Jing, W.Q.; Guan, H.J. Energy consumption prediction model for office buildings based on improved GWO-BP. Build. Energy Effic. 2022, 8, 125–129+149. [Google Scholar]
  33. Guo, C.; Shi, Y.B. Research on the Dynamic Model of Accelerometers Based on GWO-BP Method. Mea Control. Tech. 2023, 1–6. [Google Scholar]
  34. Li, Z. A Local Opposition-Learning Golden-Sine Grey Wolf Optimization Algorithm for Feature Selection in Data Classifica tion. Appl. Soft Comput. 2023, 142, 110319. [Google Scholar] [CrossRef]
  35. Pan, H.; Chen, S.; Xiong, H. A High-Dimensional Feature Selection Method Based on Modified Gray Wolf Optimization. Appl. Soft Comput. 2023, 135, 110031. [Google Scholar] [CrossRef]
  36. Pan, T.; Wang, Z.; Tao, J.; Zhang, H. Operating Strategy for Grid-Connected Solar-Wind-Battery Hybrid Systems Using Im-proved Grey Wolf Optimization. Electr. Power Syst. Res. 2023, 220, 109346. [Google Scholar] [CrossRef]
  37. Zhao, X.; Chen, Y.; Wei, G. A comprehensive compensation method for piezoresistive pressure sensor based on surface fitting and improved grey wolf algorithm. Measurement 2023, 207, 112387. [Google Scholar] [CrossRef]
  38. Sun, X.; Lei, Y. Research on financial early warning of mining listed companies based on BP neural network model. Resour. Policy 2021, 73, 102223. [Google Scholar] [CrossRef]
  39. Yesilbudak, M. Parameter extraction of photovoltaic cells and modules using grey wolf optimizer with dimension learning-based hunting search strategy. Energies 2021, 18, 5735. [Google Scholar] [CrossRef]
Figure 1. Level of gray wolves.
Figure 1. Level of gray wolves.
Electronics 12 02520 g001
Figure 2. The diagram of hunting.
Figure 2. The diagram of hunting.
Electronics 12 02520 g002
Figure 3. The diagram of attacking prey.
Figure 3. The diagram of attacking prey.
Electronics 12 02520 g003
Figure 4. BPNN structure diagram.
Figure 4. BPNN structure diagram.
Electronics 12 02520 g004
Figure 5. Flow chart of DLH-GWO-BPNN model.
Figure 5. Flow chart of DLH-GWO-BPNN model.
Electronics 12 02520 g005
Figure 6. Analysis diagram of the number of neurons.
Figure 6. Analysis diagram of the number of neurons.
Electronics 12 02520 g006
Figure 7. Hit graph of WOA-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Figure 7. Hit graph of WOA-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Electronics 12 02520 g007
Figure 8. Hit graph of GA-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Figure 8. Hit graph of GA-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Electronics 12 02520 g008
Figure 9. Hit graph of PSO-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Figure 9. Hit graph of PSO-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Electronics 12 02520 g009
Figure 10. Hit graph of GWO-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Figure 10. Hit graph of GWO-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Electronics 12 02520 g010
Figure 11. Hit graph of DLH-GWO-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Figure 11. Hit graph of DLH-GWO-BPNN: (a) prediction hit graph of the training set; (b) prediction hit graph of the test set.
Electronics 12 02520 g011
Figure 12. ROC curve comparison of PSO-BPNN, GA-BPNN, WOA-BPNN: (a) ROC curve in the training set; (b) ROC curve in the test set.
Figure 12. ROC curve comparison of PSO-BPNN, GA-BPNN, WOA-BPNN: (a) ROC curve in the training set; (b) ROC curve in the test set.
Electronics 12 02520 g012
Figure 13. ROC curve comparison of GWO-BPNN and DLH-GWO-BPNN: (a) ROC curve in the training set; (b) ROC curve in the test set.
Figure 13. ROC curve comparison of GWO-BPNN and DLH-GWO-BPNN: (a) ROC curve in the training set; (b) ROC curve in the test set.
Electronics 12 02520 g013
Table 1. Matching labels.
Table 1. Matching labels.
Matching ProbabilityMatching Labels
p ≤ 0.450 (mismatch)
0.45 < p < 0.651 (weak match)
0.65 < p < 0.852 (match)
p ≥ 0.853 (strong match)
Table 2. Dataset after mixed sampling.
Table 2. Dataset after mixed sampling.
Serial Number σ C1C2HrRpLabels
111,8181.6881−2.562216,4731.39396.30470.01793
26964.5410.349015−2.52024521.2120.6475126.385970.7591211
34250.20.0838−3.52893712.80.87366.62150.56363
414,6256.0878−0.935921,3691.46125.9963−0.10023
56414.2560.482088−2.721814106.5840.6382226.5614620.7620741
……
1611,286.07−0.57358−2.584397393.1750.6600166.0187780.721692
1710,78910.3652−0.881112,1131.12286.55290.22120
1816,590−0.0284−2.225981290.496.09140.85452
1914,220.523.797962−1.3478315,780.61.110816.5723540.3962892
20650.56.8366−4.9311708.241.08886.44280.32113
292216.87.8613−0.47682469.91.11426.86910.30560
309087.2−0.6742−2.63886393.10.70355.96210.65413
312320.0417.832956−0.473232440.8051.1226476.8416940.3038170
3211,340.892.712972−2.296614,995.351.3205886.0814610.0870382
3318062.9841−1.62981785.980.98896.28460.47351
427508.80.2174−2.32084931.30.65676.21240.75621
4310,605.211.192791−2.7284512,728.041.2037636.0136340.1994672
442391.3217.815665−0.482722530.7741.12496.8376740.3020560
4513,2985.5214−1.189617,8681.34366.24630.06532
4624,4945.89−0.863632,5671.32966.5670.0142
4724,4945.89−0.863632,5671.32966.5670.0142
542274.5257.845452−0.474812453.6321.1189236.8537770.3046030
559720.7714.248036−1.538449295.3260.9402996.1206850.502311
562326.87.8311−0.4732438.91.12326.83990.30370
572296.227.839496−0.474062447.5181.1206986.8480180.3042280
6311,2480.3364−2.643912,3661.09945.72030.29953
652216.87.8613−0.47682469.91.11426.86910.30560
666967.44.9571−1.45966196.50.88945.96010.52651
6719,3841.9381−1.775524,8941.28426.55170.11743
682216.87.8613−0.47682469.91.11426.86910.30560
Table 3. Input-hidden layer activation function and average accuracy.
Table 3. Input-hidden layer activation function and average accuracy.
Activation FunctionAccuracy of the Training Set (%)Accuracy of the Test Set (%)
tansig94.2692.38
radbas93.8389.05
softmax83.4088.10
purelin87.4586.67
Table 4. Model parameter values.
Table 4. Model parameter values.
PSO-BPNNGA-BPNNWOA-BPNNGWO-BPNNDLH-GWO-BPNN
n1010101010
a1tansigtansigtansigtansigtansig
a2softmaxsoftmaxsoftmaxsoftmaxsoftmax
N4040403030
p1×0.4×××
p2×0.1×××
C20××××
w0.6××××
c10.99××××
c22.0××××
c32.0××××
v13.55××××
v2−3.55××××
Note: n represents the number of hidden layer nodes; a1 represents the activation function between the input layer and the hidden layer; a2 represents the activation function between the hidden layer and the output layer; N represents the population size; p1 represents the probability of crossing; p2 represents the probability of variation; C represents the number of particles; w represents inertia weight; c1 represents the coefficient of inertia weight reduction; c2 represents the individual learning coefficient; c3 represents the group learning coefficient; v1 represents the upper limit of speed; v2 represents the lower limit of speed.
Table 5. Accuracy of different classification models.
Table 5. Accuracy of different classification models.
DatasetsPSO-BPNNGA-BPNNWOA-BPNNGWO-BPNNDLH-GWO-BPNN
Training set (%)89.3687.2374.4791.4991.49
Test set (%)80.9585.7180.9585.7195.24
Table 6. Recall of different classification models.
Table 6. Recall of different classification models.
DatasetsLabelsPSO-
BPNN
GA-
BPNN
WOA-BPNNGWO-BPNNDLH-GWO-BPNN
Training set01.001.000.801.001.00
11.000.910.641.001.00
20.880.940.880.940.81
30.700.600.600.700.90
Test set01.001.001.001.001.00
11.001.001.001.000.75
21.001.001.001.001.00
30.200.400.200.401.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, X.; Liu, J.; Men, H.; Ren, S.; Guo, L. Study on Downhole Geomagnetic Suitability Problems Based on Improved Back Propagation Neural Network. Electronics 2023, 12, 2520. https://doi.org/10.3390/electronics12112520

AMA Style

Zhou X, Liu J, Men H, Ren S, Guo L. Study on Downhole Geomagnetic Suitability Problems Based on Improved Back Propagation Neural Network. Electronics. 2023; 12(11):2520. https://doi.org/10.3390/electronics12112520

Chicago/Turabian Style

Zhou, Xu, Jing Liu, Huiwen Men, Shangsheng Ren, and Liwen Guo. 2023. "Study on Downhole Geomagnetic Suitability Problems Based on Improved Back Propagation Neural Network" Electronics 12, no. 11: 2520. https://doi.org/10.3390/electronics12112520

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop