Next Article in Journal
Nanomodified Concrete with Enhanced Characteristics Based on River Snail Shell Powder
Previous Article in Journal
NDT Method for Line Laser Welding Based on Deep Learning and One-Dimensional Time-Series Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Missing Well Logs Prediction Based on Hybrid Kernel Extreme Learning Machine Optimized by Bayesian Optimization

1
Hebei Instrument & Meter Engineering Technology Research Center, Hebei Petroleum University of Technology, Chengde 067000, China
2
State Key Laboratory of Nuclear Resources and Environment, East China University of Technology, Nanchang 330013, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2022, 12(15), 7838; https://doi.org/10.3390/app12157838
Submission received: 4 July 2022 / Revised: 1 August 2022 / Accepted: 1 August 2022 / Published: 4 August 2022

Abstract

:
Geophysical logging is an essential measurement tool in the oil/gas exploration and development field. In practice, predicting missing well logs is an effective way to reduce the exploration expenses. Because of the complexity and heterogeneity of the reservoir, there must be strong nonlinear correlations between the well logs. To improve the accuracy and stability of the missing well logs prediction, a method based on a Bayesian optimized hybrid kernel extreme learning machine (BO-HKELM) algorithm is proposed. Firstly, the LightGBM algorithm is applied to screen out important features related to the missing well logs and reduce the input dimension of the prediction model. Secondly, the hybrid kernel extreme learning machine (HKELM) algorithm is applied to construct the missing well logs prediction model, and the hyperparameters ( C 0 , C 1 , d , σ , C ) of the model are optimized by the Bayesian algorithm. Finally, the BO-HKELM model is applied to the prediction of the missing well logs in a block of the Ordos Basin in China. The results show that the RMSE, MAE, and R-square predicted by the BO-HKELM model are 0.0767, 0.0613, and 0.9029, respectively. It can be found that the BO-HKELM model has better regression accuracy and generalization ability, and can estimate missing logs more accurately than the traditional machine learning methods, which provides a promised method for missing well logs estimation.

1. Introduction

In oil/gas exploration, geophysical logging is an effective way of analyzing and describing the subsurface conditions by using physical measurements. The primary measuring items are the electrochemical, acoustic, radioactive, and other geophysical features of the rock formation. In reservoir evaluation, the lithology, physical property, and oil-gas bearing properties of the reservoir can be comprehensively evaluated by using the results of logging data analysis and interpretation. It is essential to merge the various logs and seismic data in order to efficiently develop a reservoir model and decrease the fuzziness of geological interpretation. However, when the logging instrument is faulty, the borehole wall is damaged and collapses occur, which will lead to the loss of logging curves in the working area [1].
It is of great significance to use the logging curves of known intervals to supplement the missing logging curves for subsequent reservoir prediction. Therefore, numerous studies have been carried out in an effort to identify the non-linear correlations between various well logs. According to the investigation, the most frequently used approaches include the empirical model [2,3], the multi-variable linear regression model [4,5] and the traditional machine learning model [6,7]. The empirical model can result in the major deviations because of the strong nonlinear relationship between the reservoir characteristics and missing well logs. The multi-variable linear regression model has the disadvantage of inadequate equation expression. The traditional machine learning model has issues with over-fitting or being trapped in local optimums, as well as the need for sufficient and compact training sample sets [8,9,10]. Therefore, a missing well logs prediction model with excellent computation speed, prediction accuracy and generalization ability is urgently needed.
The extreme learning machine (ELM) is a single hidden layer feedforward neural network with the characteristics of a simple structure and fast learning speed. However, the input weights and the threshold of the hidden layer of the ELM are randomly set, and the appropriate number of the hidden layer is difficult to determine. The kernel extreme learning machine (KELM) uses kernel mapping instead of random mapping, which greatly reduces the complexity of the network and means that the model has better prediction and generalization ability. However, the KELM usually adopts a single kernel function in the application process, which is difficult to adapt to samples with multiple data characteristics [11].
The hybrid kernel extreme learning machine (HKELM) is built by weighing multiple kernel functions to enhance the regression accuracy, which can solve the problem that it is difficult for the KELM with a single kernel function is to obtain the characteristics of a multidimensional sample. However, the parameters of the HKELM model are stochastic, and the training accuracy and time are easily affected by randomness. Therefore, it is necessary to use the optimization algorithm to optimize the hyper-parameters of the HKELM model [12,13,14,15,16]. It has been established that the Bayes optimization algorithm offers exceptional search performance [17]. In this study, a hybrid kernel extreme learning machine, optimized by a Bayesian optimization algorithm (BO-HKELM), is proposed to predict the missing well logs. Furthermore, the BO-HKELM model is applied to two wells (Well A1 and Well A2) in a block of the Ordos Basin in China. It can be found that the BO-HKELM model has higher precision compared to the other models proposed in this research.

2. Principle and Modeling

2.1. Principle of KELM

The ELM is a practical and efficient feedforward neural network. The structure of the ELM network is shown in Figure 1.
In the diagram, X , Y denote the input and output dataset, respectively.   x k = x k 1 x k 2 x k n T is the input vector at dimension n , and y k = y k 1 y k 2 y k l T is the output vector at dimension l . w j = w 1 j w 2 j w n j is the input weight of the node j in the hidden layer, and β j = β j 1 β j 2 β j l T is the output weight of the node j in the same hidden layer. b = b 1 b 2 b m T is the threshold of the hidden layer. g x = 1 1 + e x is the excitation function of the hidden layer. f ELM x k is the network output of the ELM, which is expressed as follows:
f ELM x k = j = 1 m β j · g w j x k + b j .
The output matrix f ELM X can be written as f ELM X = H β , where H is the output matrix of the hidden layer:
H = g w 1 · x 1 + b 1 g w m · x 1 + b m g w 1 · x 2 + b 1 g w m · x 2 + b m g w 1 · x N + b 1 g w m · x N + b m N × m .
X and Y are both known variables. The input weight matrix W and the threshold b can be randomly given. H can be calculated by Equation (2). The output weight matrix β is calculated by the formula β = H + · Y , where H + is the pseudo-inverse matrix of H .
The ELM overcomes the shortcomings of the traditional neural network, such as slow training speed, the overfitting that is prone to occur, and falling into local extremums, but there are still problems, such as difficulty in determining the number of nodes in the hidden layer and the occurrence of overfitting. To address the shortcomings of the ELM, inspired by the kernel function introduced into the support vector machines (SVM), the KELM is proposed in the literature [18], which denotes the mapping from the input to the output of the hidden layer as h X .
The kernel matrix is defined as
Ω K E L M X = h X · H T .
The elements at row i and column j of the matrix are shown by the following equation:
Ω K E L M i j x i , x j = h x i · h x j K x i · x j
where K x i · x j is the kernel function. f K E L M X can be estimated by Formula (5), which is described below.
f K E L M X = h X · H T Ω K E L M X 1 · Y = K X · x 1   K X · x N × Ω K E L M X 1 · Y .
After introducing the regularized item, the following equation is used:
f K E L M X = K X · x 1     K X · x N × I C + Ω K E L M X 1 · Y .
Then, the network output f KELM X k   of the KELM can be written as
f K E L M X k = k = 1 n Ω K E L M X · x k · ( ( I C ) k + Ω K E L M X · x k ) 1 · Y k
where C is the regular term coefficient. The accuracy of the model increases as the parameter C increases, but overfitting is more likely to happen. Conversely, as C decreases, the generalization ability increases and error-tolerant rate increases, but underfitting is more likely to happen.

2.2. Hybrid Kernel Extreme Learning Machine

By using the kernel function to map the low-dimensional space to the high-dimensional space, the KELM greatly reduces the complexity of the network, and makes the prediction and the generalization ability better [19]. However, even within the same sample, the performance of various kernel functions for prediction is extremely diverse [20], indicating that it is difficult for a single kernel function in the standard KELM algorithm to accommodate diverse logging sample data. Therefore, a hybrid kernel extreme learning machine (HKELM) is proposed. By adding a hybrid kernel, the defect of ELM with a single core can be overcome, and the problem of low prediction accuracy and insufficient generalization ability can be solved.
Based on the synthesis of the characteristics of common kernels and the trade-off between model accuracy and computational complexity, two kinds of kernel functions, polynomial and Gaussian radial basis, are selected to carry out a weighted combination, and the equivalent kernel function, combining the two kinds of kernels, is constructed. The expressions of the two basic kernel functions are shown in Equations (8) and (9), respectively.
Polynomial kernel function can be expressed as
K poly X , x k = x k T X + C 0 d , C 0 > 0
where C 0 and d are the parameters of the polynomial kernel function. This kernel function can generally represent the nonlinear mapping of the system.
Gaussian radial basis kernel function can be expressed as
K R B F X , x k = e x p X x k 2 2 σ 2
where σ is the kernel width. This kernel function can generally represent the nonlinearity of the system.
Therefore, the equivalent kernel function used in HKELM algorithm can be expressed as the following equation:
Ω M K E L M X , x k = C 1 K poly X , x k + C 2 K R B F X , x k
where C 1 and C 2 are the weighted coefficients of the kernel function, ranging between (0, 1) and   C 1 + C 2 = 1 . After cancelling out C 2 ,
Ω MKELM X , x k = C 1 · K poly X , x k + 1 C 1 · K R B F X , x k .
After substituting Equation (11) into Equation (7), the output f M K E L M x k of the proposed HKELM algorithm that integrates the two kernel functions is expressed as
f M K E L M x k = k = 1 N Ω M K E L M X · x k C 1 x k T X + C 0 d + 1 C 1 e x p X x k 2 2 σ 2 + I C k 1 · Y k .
The benefits of the global kernel and the local kernel can be combined with the hybrid kernel function of HKELM. So, the HKELM not only has good local search ability but also strengthens the global search ability. Due to the inefficiency in manually determining the hyper-parameter, the HKELM model shoule be optimized by the optimization algorithm.

2.3. Bayesian Optimization Algorithm

The Bayesian optimization algorithm is an intelligent optimization algorithm that can effectively reflect the relationship between variables [21]. The principle of Bayesian optimization is to use Bayes’ theorem to estimate the posterior distribution of the objective function and then select the next combination of hyperparameters to be sampled according to the distribution. This method makes full use of the sampling results of the previous sampling points to improve the shape of the objective function and find the global optimal hyperparameter. One must let X = x 1 , x 2 , , x n be a set of hyperparameter combinations and f x be the objective function of the hyperparameter x . The principle of Bayesian optimization is to find x , which can be written as
x = a r g m a x f x .
The optimization method consists of two parts, a probabilistic surrogate model and a sampling function.
According to Bayes’ theorem, the model parameters are updated in the following equation:
p f | D 1 : t = p D 1 : t | f p f p D 1 : t
where D 1 : t = x 1 , f x 1 , x 2 , f x 2 , , x t , f x t is the sample set; p D 1 : t | f is the likelihood distribution of y ; p f is the prior probability model of f ; p D 1 : t is the marginal likelihood distribution, and p f | D 1 : t is the posterior probability model of f .
The sampling function is the referenced credential for the Bayesian optimization method to obtain the next sample point in the hyperparameter space. In this research, expected improvement (EI) is used as the sampling function, which is expressed as
α x = μ x q + Z + σ x φ z x t + 1 = a r g m a x f x
where α x is the objective function; μ x is the mean value; σ x is the standard deviation value; q + is the maximum value of the current objective function; Z is the cumulative distribution function of the Gaussian distribution; φ z is the probability density function of the distribution, and x t + 1 is the hyperparameter of the evaluation.

3. Process of Predicting Missing Logging Curves Based on BO-HKELM

In this research, the specific steps for optimizing the parameters ( C 0 , C 1 , d , σ , C ) in HKELM by the Bayesian algorithm are as follows.
In Step 1, one must divide the input logging dataset into a training set and test set, and normalize the data to (0, 1) using the max–min normalization method. The expression of the max–min normalization method is as follows:
x = x x m i n x m a x x m i n
where x is the actual vector. x m a x and x m i n are the maximum and minimum values of the vector x, respectively. x is the normalized vector.
In Step 2, one must initialize the model, including the value range, the initial population, the maximum iterations, and the upper accuracy setting of the parameters ( C 0 , C 1 , d , σ , C ) in HKELM.
In Step 3, one must randomly select a group from the initial population as the initial solution.
In Step 4, one must construct the probability distribution function of the Bayesian network through the Gaussian process, and establish or update an objective function model. The objective function is the root mean square error (RMSE), which is expressed as
R M S E = 1 N i = 1 N y k y ^ k 2
where y k is the true value, and y ^ k is the predicted value of HKELM.
In Step 5, one must determine the position of the next cycle parameter combination according to the acquisition function.
In Step 6, one must calculate the probability value of the parameter combination and update the population distribution.
In Step 7, one must determine whether the resulting combination of parameters satisfies the condition; if so, output the optimum results of C 0 , C 1 , d , σ , C . If not, one must return to Step 4 and repeat the experiment until the condition is satisfied or the maximum number of iterations is reached.
In Step 8, one must apply the optimum parameter combination ( C 0 , C 1 ,   d , σ , C ) to the missing well logs estimation of the HKELM.
In Step 9, one must evaluate the accuracy of the optimized prediction results using the test sample set, and compare the changes in accuracy before and after optimization.
The optimization process is shown in Figure 2.

4. Practical Application and Result Analysis

4.1. Data Preparation

In this research, two exploration wells located in a block of the Ordos Basin in China are studied. The Ordos Basin has the character of stable depression tectonic basin, with the metamorphic crystalline rock series of Archaea and Middle Proterozoic as the rigid sedimentary basement, overlying the platform type sedimentary caps of Middle and Upper Proterozoic, Early Paleozoic, Late Paleozoic and Mesozoic. The Mesozoic is a single fluvial and lacustrine facies terrine detrital coal-bearing sedimentary formation, characterized by the large thickness and wide distribution of lacustrine deltas in the Triassic Yanchang Formation, which is one of the most important oil-gas enrichment horizons in Ordos Basin.
We designate the two wells as Well A1 and Well A2, respectively, to maintain the privacy of well information. Additionally, since the petrophysical logs are primarily found in the intervals of the oil/gas reservoirs, the interest interval of the two selected wells is focused on the Upper Triassic Yanchang Formation. The log graphs of Well A1 and Well A2 are shown in Figure 3 and Figure 4, respectively. The petrophysical logs in the studied interval of the two wells are DEPTH, caliper (CAL), neutron porosity (NPHI), density (RHOB), gamma ray (GR), acoustic time difference (DT), and true resistivity (RT). In addition, DEPTH, CAL, NPHI, RHOB, GR, DT, and RT can be used to measure depth, diameter, porosity, density, shale content, acoustic time difference, and true resistivity of the borehole surrounding rock, respectively. The response of the logging sequence is a comprehensive reflection of the lithology, physical property, electrical property and oil-gas property of the corresponding underground formation. Caliper logging is used to indicate hole expansion and aid in lithology identification and borehole correction. Generally, the oil and gas have the physical characteristics of lower density, higher acoustic time difference and higher resistivity compared with water. In a conventional reservoir, oil and gas are usually filled in the pores and fractures of the formation. Meanwhile, the lower the mud content in the formation, the higher the porosity. Therefore, in the conventional reservoir, the logging response characteristics of high porosity, high acoustic time difference, high resistivity, low mud content and low density may indicate hydrocarbon enrichment. In this research, we suppose the RHOB log is missing and forecast the “missing” RHOB curve using other logging data by BO-HKELM.
The statistical properties of the two selected wells are presented in Table 1. There are 1601 and 1202 data points in Well A1 and Well A2, respectively.

4.2. Feature Analysis

In this research, the decision tree-based distributed gradient lifting framework (LightGBM) is used to assess the importance of the features. LightGBM is a lightweight framework of the gradient boosting decision tree algorithm. It uses decision tree iterative training to obtain the optimal model, which has the advantages of excellent training effects and difficulty to over-fit. Compared with the extreme gradient boosting, LightGBM uses a gradient-based one-side sampling algorithm to accelerate the training speed of the model, without compromising accuracy [22].
Six parameters, including DEPTH, CAL, NPHI, GR, DT, and RT, from Well A1 are input into the LightGBM model to obtain the influence of each parameter on RHOB and order the characteristic importance. The quantitative results are shown in Table 2:
The features are added one by one in descending order of relevance to the ordinary model (KELM) in order to choose the significant sample features. Figure 5 displays the change curve of model errors (RMSE) that is determined by the comparison test. The declining trend of RMSE shows a substantial change when the model contains four inputs, as shown in Figure 5. Therefore, to reduce the computational complexity and the time duration, a model with four inputs is applied to estimate the “missing” RHOB logs. The four input variables are DT, NPHI, CAL, and GR.

4.3. Optimization of Algorithm Parameters

In this research, it is assumed that Well A1 in the studied interval has the complete RHOB curve, and the RHOB curve of Well A1 is used as the training set to establish the ELM, KELM, HKELM, and BO-HKELM models. The “missing” RHOB curve of Well A2 in the corresponding interval is predicted.
The Bayesian optimization algorithm introduced in Section 2.3 is used to optimize five parameters in the BO-HKELM model, such as the polynomial kernel function parameters C 0 and d , the Gaussian radial basis kernel function width σ , the kernel weighting coefficient C 1 , and the the regular term coefficient C . These five tuning parameters control the developed HKELM’s capacity for prediction, learning, and generalization. To acquire the optimized hyperparameters of the HKELM model, all of the values for each parameter ( C 0 , C 1 , d , σ , C ) are placed in five distinct vectors. The goal of the optimization procedure is to compare the different parameter combinations and select the optimal ( C 0 , C 1 , d , σ , C ) values with the lowest cost function. The fitness (RMSE) reduction rate in the training phase of HKELM in Well A1 is shown in Figure 6.
It can be observed that the fitness value decreases sharply in the optimization process of the first 53 generations, and the best fitness value is found around the 50th generation, and then the fitness value remains basically unchanged. The parameter optimization results of the Bayesian algorithm are shown in Table 3.

4.4. Comparative Analysis of Model Prediction Effect

The quantitative prediction results are shown in Table 4. It can be observed that the KELM (radial basis function, RBF) model has better prediction accuracy than the ELM model, indicating that the addition of the kernel function can improve the prediction effect. We can also observe that the prediction accuracy of the HKELM model is better than that of KELM (RBF) model, which indicates that the addition of a hybrid kernel function can also improve the prediction ability. The quantitative prediction results show the higher accuracy of the BO-HKELM model compared to the ELM, KELM (RBF) and HKELM models. It can be interpreted that the BO-HKELM model adopts the hybrid kernel and the Bayesian optimization algorithm to obtain the optimal parameters, which improves the generalization ability and prediction accuracy.
The comparison between measured and estimated RHOB in Well A2 by the trained models (ELM, KELM (RBF), HKELM, and BO-HKELM) in Well A1 is shown in Figure 7. It can be observed that when the change in the logging curve is relatively stable, the output results of the BO-HKELM model and other models show little difference. Meanwhile, when the log curve has a local abrupt change (the green frame), the prediction result of the BO-HKELM model is closer to the measured value than the other models, indicating that the BO-HKELM model can extract the spatial characteristics of the log curve more. Therefore, compared with the ordinary machine learning model, the BO-HKELM model is more suitable for missing logging curve prediction.
The cross-plots of original RHOB vs predicted RHOB using ELM, KELM (RBF), HKELM, and BO-HKELM in Well A2 are shown in Figure 8. It can be observed that the predicted results of BO-HKELM are closer to the actual RHOB than the other models, indicating that the BO-HKELM is better than other models in fitting multiple local details of DTS. In addition, it can be observed from the quantitative prediction results in Table 4 that compared with other models, the BO-HKELM model has lower errors and higher accuracy, which also reflects its advantages.
To compare the prediction ability of the BO-HKELM model and other models in the case of local abrupt changes in the logging curves, the prediction results of Well A2 are locally amplified and analyzed. The comparison outputs are shown in Figure 9. The black curve is the measured RHOB, and the red, orange, purple and green curves represent the prediction results of BO-HKELM, HKELM, KELM and ELM, respectively. It can be found from the Figure 9 that when the logging curve smoothly changes, the output results of the BO-HKELM model and other models show little difference. While the log curve has the local mutation, the prediction result of the BO-HKELM model is closer to the measured value than other models. The reason is that the BO-HKELM model adds a hybrid kernel function and adopts a Bayesian algorithm to optimize the hyper-parameters, which can fully extract the spatial characteristics of the log curve and enhance the prediction ability of the model.
In conclusion, the BO-HKELM model proposed in this research has an excellent prediction effect overall and reveals a local abrupt change in the logging curves, clearly demonstrating its advantages in predicting the missing well logs.

5. Conclusions

A missing well logs prediction method based on the hybrid BO-HKELM algorithm is proposed in this research by using the common petrophysical data from a field of the Ordos Basin in China. The ELM, KELM (RBF), and HKELM algorithms are used to assess the precision and generalizability of the obtained model. The LightGBM algorithm is applied in conjunction with the original model (HKELM) to show that modeling errors decrease as the number of estimator model inputs increase; however, the trend is only minimal after four input variables. Thus, it can be concluded that the variables DT, NPHI, CAL, and GR are the appropriate modeling inputs. In comparison to the other models studied in this research, the BO-HKELM model has the highest accuracy (R-square: 0.9029) and the lowest error (RMSE: 0.0767 and MAE: 0.0613). In addition, the BO-HKELM model outperforms other models in terms of fitting different local logging curve features, and it is better at rebuilding local curve characteristics when the reservoir curve exhibits a local mutation. The reason is that the BO-HKELM model incorporates a hybrid kernel function and uses a Bayesian algorithm to optimize the parameter weights, which may extract the spatial properties of the log curve more effectively and improve the model’s predictive ability. Therefore, it can be confidently stated that the BO-HKELM model has higher precision and is much better suited for missing well logs prediction, compared to the other models studied in this research.

Author Contributions

Conceptualization, L.Q. and Y.C.; methodology, L.Q.; software, L.Q.; validation, L.Q., Y.C. and Z.J.; formal analysis, H.S.; investigation, K.X.; resources, K.X.; data curation, K.X.; writing—original draft preparation, L.Q.; writing—review and editing, L.Q.; visualization, H.S.; supervision, Z.J.; project administration, Z.J.; funding acquisition, K.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Key Research and Development Program of China (2016YFC0600201), the Academic and technical leader Training Program of Jiangxi Province (20204BCJ23027) and the Joint Innovation Fund of State Key Laboratory of Nuclear Resources and Environment (2022NRE-LH-18).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to the confidentiality requirements of the data provider.

Acknowledgments

China petrochemical corporation (Sinopec) is gratefully acknowledged for providing useful data and valuable support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sun, Y.Z.; Huang, J. Application of multi-task deep learning in reservoir shear wave prediction. Prog. Geophys. 2021, 36, 799–809. [Google Scholar]
  2. Gardner, G.; Gardner, L.W.; Gregory, A.R. Formation velocity and density; the diagnostic basics for stratigraphic traps. Geophysics 1974, 39, 770–780. [Google Scholar] [CrossRef]
  3. Smith, J.H. A Method for Calculating Pseudo Sonics from E-Logs in a Clastic Geologic Setting. Gcags Transactions 2007, 57, 1–4. [Google Scholar]
  4. Wang, J.; Liang, L.; Qiang, D.; Tian, P.; Tan, W. Research and application of reconstructing logging curve based on multi-source regression model. Lithologic Reserv. 2016, 28, 113–120. [Google Scholar]
  5. Liao, H.M. Multivariate regression method for correcting the influence of expanding diameter on acoustic curve of density curve. Geophys. Geochem. Explor. 2014, 38, 174–179. [Google Scholar]
  6. Banchs, R.; Jiménez, J.R.; Pino, E.D. Nonlinear estimation of missing logs from existing well log data. In Proceedings of the 2001 SEG Annual Meeting, San Antonio, TX, USA, 2–6 September 2001. [Google Scholar]
  7. Salehi, M.M.; Rahmati, M.; Karimnezhad, M.; Omidvar, P. Estimation of the non records logs from existing logs using artificial neural networks. Egypt. J. Pet. 2016, 26, 957–968. [Google Scholar] [CrossRef]
  8. Ramachandram, D.; Taylor, G.W. Deep Multimodal Learning: A Survey on Recent Advances and Trends. IEEE Signal Process. Mag. 2017, 34, 96–108. [Google Scholar] [CrossRef]
  9. Zhang, Q.; Yang, L.T.; Chen, Z.; Li, P. A survey on deep learning for big data. Inf. Fusion 2018, 42, 146–157. [Google Scholar] [CrossRef]
  10. Jian, H.; Chenghui, L.; Zhimin, C.; Haiwei, M. Integration of deep neural networks and ensemble learning machines for missing well logs estimation. Flow Meas. Instrum. 2020, 73, 101748. [Google Scholar] [CrossRef]
  11. Lin, T.; Cai, R.; Zhang, L.; Yang, X.; Liu, G.; Liao, W. Prediction intervals forecasts of wind power based on IBA-KELM. Renew. Energy Resour. 2018, 36, 1092–1097. [Google Scholar]
  12. Abdullah, M.; Rashid, T.A. Fitness Dependent Optimizer: Inspired by the Bee Swarming Reproductive Process. IEEE Access 2019, 7, 43473–43486. [Google Scholar] [CrossRef]
  13. Cheng, R.; Jin, Y. A Competitive Swarm Optimizer for Large Scale Optimization. IEEE Trans. Cybern. 2015, 45, 191–204. [Google Scholar] [CrossRef] [PubMed]
  14. Zhao, W.; Wang, L.; Zhang, Z. Supply-Demand-Based Optimization: A Novel Economics-Inspired Algorithm for Global Optimization. IEEE Access 2019, 7, 73182–73206. [Google Scholar] [CrossRef]
  15. Shabani, A.; Asgarian, B.; Salido, M.A.; Gharebaghi, S.A. Search and Rescue optimization algorithm: A new optimization method for solving constrained engineering optimization problems. Expert Syst. Appl. 2020, 161, 113698. [Google Scholar] [CrossRef]
  16. Das, B.; Mukherjee, V.; Das, D. Student psychology based optimization algorithm: A new population based optimization algorithm for solving optimization problems—ScienceDirect. Adv. Eng. Softw. 2020, 146, 102804. [Google Scholar] [CrossRef]
  17. Sultana, N.; Hossain, S.; Almuhaini, S.H.; Düştegör, D. Bayesian Optimization Algorithm-Based Statistical and Machine Learning Approaches for Forecasting Short-Term Electricity Demand. Energies 2022, 15, 3425. [Google Scholar] [CrossRef]
  18. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme Learning Machine for Regression and Multiclass Classification. IEEE Trans. Syst. Man Cybern. Part B 2012, 42, 513–529. [Google Scholar] [CrossRef] [PubMed]
  19. Heidari, A.A.; Rahim, A.A.; Chen, H. Efficient boosted grey wolf optimizers for global search and kernel extreme learning machine training. Appl. Soft Comput. 2019, 81, 105521. [Google Scholar] [CrossRef]
  20. Chao, Z.; Yin, K.; Cao, Y.; Intrieri, E.; Ahmed, B.; Catani, F. Displacement prediction of step-like landslide by applying a novel kernel extreme learning machine method. Landslides 2018, 15, 2211–2225. [Google Scholar]
  21. Pelikan, M. Bayesian Optimization Algorithm: From Single Level to Hierarchy. Master’s Thesis, University of Illinois at Urbana-Champaign, Champaign, IL, USA, 2002. [Google Scholar]
  22. Wang, X.; Zhang, G.; Lou, S.; Liang, S.; Sheng, X. Two-round feature selection combining with LightGBM classifier for disturbance event recognition in phase-sensitive OTDR system. Infrared Phys. Technol. 2022, 123, 104191. [Google Scholar] [CrossRef]
Figure 1. ELM model structure diagram.
Figure 1. ELM model structure diagram.
Applsci 12 07838 g001
Figure 2. Flow chart of optimizing parameters of HKEM based on Bayesian algorithm.
Figure 2. Flow chart of optimizing parameters of HKEM based on Bayesian algorithm.
Applsci 12 07838 g002
Figure 3. Well logging graph of Well A1 in the studied interval.
Figure 3. Well logging graph of Well A1 in the studied interval.
Applsci 12 07838 g003
Figure 4. Well logging graph of Well A2 in the studied interval.
Figure 4. Well logging graph of Well A2 in the studied interval.
Applsci 12 07838 g004
Figure 5. RMSE variation curve of different characteristic numbers.
Figure 5. RMSE variation curve of different characteristic numbers.
Applsci 12 07838 g005
Figure 6. The fitness reduction rate in different iterations of Bayesian optimization algorithm for training HKELM.
Figure 6. The fitness reduction rate in different iterations of Bayesian optimization algorithm for training HKELM.
Applsci 12 07838 g006
Figure 7. Visual results of the RHOB prediction results of Well A2.
Figure 7. Visual results of the RHOB prediction results of Well A2.
Applsci 12 07838 g007
Figure 8. (a) BO-HKELM, (b) HKELM, (c) KELM, (d) ELM. The cross plot of measured RHOB versus the predicted one using BO-HKELM, HKELM, KELM, and ELM at testing phases in Well A2.
Figure 8. (a) BO-HKELM, (b) HKELM, (c) KELM, (d) ELM. The cross plot of measured RHOB versus the predicted one using BO-HKELM, HKELM, KELM, and ELM at testing phases in Well A2.
Applsci 12 07838 g008aApplsci 12 07838 g008b
Figure 9. Local enlarged display of prediction results in Well A2.
Figure 9. Local enlarged display of prediction results in Well A2.
Applsci 12 07838 g009
Table 1. Statistical characteristics of Well A1 and Well A2 in the studied interval.
Table 1. Statistical characteristics of Well A1 and Well A2 in the studied interval.
Well NameIndexPetrophysical Well Logs
DEPTH
(m)
CAL
(cm)
DT
(μs/m)
GR
(API)
NPHI
(v/v)
RT
(ohmm)
RHOB
(g/cm3)
Well A1Min1880.0021.95187.3814.970.036.411.21
25%1910.0023.11231.2953.000.1716.712.24
50%1940.0024.53237.7886.110.2027.232.31
Mean1940.0025.03250.9793.530.2730.442.29
75%1970.0326.21260.20132.760.3443.072.47
Max2000.0041.88407.97212.320.9270.842.70
Well A2Min1560.0021.36210.1215.630.075.561.30
25%1597.4622.11230.2465.500.1216.752.34
50%1635.0022.40238.1399.620.1721.782.48
Mean1635.0023.17249.1791.190.1922.732.39
75%1672.5322.87251.64114.390.2328.392.54
Max1710.0039.43400.92149.030.6989.922.70
Table 2. Importance scores of the parameters based on the LightGBM model.
Table 2. Importance scores of the parameters based on the LightGBM model.
RankFeatureScore
1DT865
2NPHI661
3CAL594
4GR466
5RT448
6DEPTH429
Table 3. The optimal parameters optimized by Bayesian algorithm in Well A1.
Table 3. The optimal parameters optimized by Bayesian algorithm in Well A1.
BO-HKELM C 0 C 1 d σ C
Trained by Well A121.430.24269.0386.03
Table 4. The outcomes of estimating RHOB in Well A2 using the trained models from Well A1.
Table 4. The outcomes of estimating RHOB in Well A2 using the trained models from Well A1.
ModelRMSEMAER-Square
ELM0.11330.07980.7916
KELM0.10550.06380.8178
HKELM0. 09010.06230.8874
BO-HKELM0.07670.06130.9029
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qiao, L.; Cui, Y.; Jia, Z.; Xiao, K.; Su, H. Missing Well Logs Prediction Based on Hybrid Kernel Extreme Learning Machine Optimized by Bayesian Optimization. Appl. Sci. 2022, 12, 7838. https://doi.org/10.3390/app12157838

AMA Style

Qiao L, Cui Y, Jia Z, Xiao K, Su H. Missing Well Logs Prediction Based on Hybrid Kernel Extreme Learning Machine Optimized by Bayesian Optimization. Applied Sciences. 2022; 12(15):7838. https://doi.org/10.3390/app12157838

Chicago/Turabian Style

Qiao, Lei, You Cui, Zhining Jia, Kun Xiao, and Haonan Su. 2022. "Missing Well Logs Prediction Based on Hybrid Kernel Extreme Learning Machine Optimized by Bayesian Optimization" Applied Sciences 12, no. 15: 7838. https://doi.org/10.3390/app12157838

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop