Next Article in Journal
Measuring Dental Enamel Thickness: Morphological and Functional Relevance of Topographic Mapping
Previous Article in Journal
Quantifying the Displacement of Data Matrix Code Modules: A Comparative Study of Different Approximation Approaches for Predictive Maintenance of Drop-on-Demand Printing Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fast Reservoir Characterization with AI-Based Lithology Prediction Using Drill Cuttings Images and Noisy Labels

1
Aramco Innovations LLC, 117127 Moscow, Russia
2
EXPEC Advanced Research Center Saudi Aramco, Dhahran 34466, Saudi Arabia
*
Author to whom correspondence should be addressed.
J. Imaging 2023, 9(7), 126; https://doi.org/10.3390/jimaging9070126
Submission received: 9 May 2023 / Revised: 6 June 2023 / Accepted: 19 June 2023 / Published: 21 June 2023
(This article belongs to the Section AI in Imaging)

Abstract

:
In this paper, we considered one of the problems that arise during drilling automation, namely the automation of lithology identification from drill cuttings images. Usually, this work is performed by experienced geologists, but this is a tedious and subjective process. Drill cuttings are the cheapest source of rock formation samples; therefore, reliable lithology prediction can greatly reduce the cost of analysis during drilling. To predict the lithology content from images of cuttings samples, we used a convolutional neural network (CNN). For training a model with an acceptable generalization ability, we applied dataset-cleaning techniques, which help to reveal bad samples, as well as samples with uncertain labels. It was shown that the model trained on a cleaned dataset performs better in terms of accuracy. Data cleaning was performed using a cross-validation technique, as well as a clustering analysis of embeddings, where it is possible to identify clusters with distinctive visual characteristics and clusters where visually similar samples of rocks are attributed to different lithologies during the labeling process.

1. Introduction

Lithology data are essential for many applied problems of petroleum engineering, including reservoir characterization, inter-well correlation, geological model construction, basin and petroleum system modeling, etc. The analysis of rock sample recovering during borehole drilling can be considered as one of the primary and direct sources of information about rock lithology. These samples are known as rock cuttings. The initial data on the lithological content of rock cuttings are obtained on-site by geologists that have relevant expertise for conducting cuttings description. More precise data on rock lithology are obtained by means of laboratory equipment (such as X-ray diffraction, X-ray fluorescence, near-infrared spectroscopy, etc.) after the drilling is completed and representative sample collection is delivered for detailed analysis in the laboratory or data, obtained during drilling [1,2,3,4].
Due to economic reasons, a common practice is to conduct detailed laboratory investigations of core samples that were recovered from the ‘sweet-spots’ that represent hydrocarbon-bearing intervals of the well. Thus, precise data on rock lithology are often constrained to some specific depth intervals. Moreover, some results of laboratory investigations may not be representative in case of a high heterogeneity of the drilled formations.
At the same time, drill cuttings offer a rich source of data that encompasses a broader stratigraphic range compared to full-sized cores. Analyzing them in near-real-time could enable cost-effective and fast reservoir characterization. Therefore, the development of an automated pipeline for an on-site description of drill cuttings lithology is a high-priority task for many petroleum companies.
In this work, we researched the possibility of detecting lithology using just white light images, which are the cheapest data to obtain during drilling. The commonly used pipeline to infer data on rock lithology from cuttings is the following. During the drilling of a wellbore, a rotating bit breaks the rocks downhole, and drilling mud carries the broken fragments to the surface. At the surface, these fragments move to a shaker, which separates the drilling mud from the rocks for recirculation. Cuttings are cleaned from the drilling mud and a field geologist conducts a manual visual description of the recovered samples based on their expertise.
The manual description of rock cuttings very often has the following problems: (1) biased descriptions due to the subjective nature of the results, (2) a time-consuming process, (3) human errors due to mud contamination of cuttings. Moreover, some types of rocks can be very close in terms of visual appearance, which complicates the process of visual description. One of the possible solutions for these problems is the implementation of an automated workflow for detecting rock lithology from rock cuttings.
Within our research, we developed and proposed a new approach for estimating the lithological composition of drill cuttings from digital photographs. The proposed method relies on the application of computer vision techniques and is data-driven. The basis for the approach development was a dataset comprising high-resolution images and results of drill cuttings description conducted by field geologists from more than 20 industrial wells. The dataset includes cuttings images for nine lithotypes, including clastic and carbonate rocks. We developed a convolutional deep learning model that predicts the likelihood of a sample belonging to a specific rock class. To understand the generalization ability of the developed model, we tested our approach on an unseen well.
Much effort has been invested in the cuttings images analysis, with some valuable research contributions that we will briefly outline. In [5], the authors suggested the application of several classification approaches to predict rock lithology from drill cuttings. The initial approach was to use a pre-trained AlexNet combined with a support vector machine classification of learned features. However, the validation results were below expectations. Thus, the authors tried to use a Bayesian-optimized network and network ensemble, which performed worse than a single network. They utilized a dataset from the 2019 EAGE hackathon, which is now unavailable. With a validation set of 34 images, they achieved an 80%+ accuracy but noted that some labels in the set might be erroneous. This highlights the challenge of accurately classifying drill cuttings containing rocks with different lithologies, as they can appear very similar and are prone to misclassification.
In [6], the authors focused on a specific case of drilling cuttings produced by polycrystalline diamond compact (PDC) bits, which yield small-sized pieces. The paper outlines a segmentation method for determining the quantity of oil-bearing cuttings in an image using a special type of fluorescent light. The segmentation was carried out using the watershed algorithm, and lithology was identified by comparing image features with known features from a feature library. However, this algorithm cannot provide the lithological composition of cuttings extracted from a well during drilling. Instead, it suggests comparing feature vectors with those in the available feature library of image areas. In [7], the authors proposed using data on measurements of elemental rock composition alongside images, but such data require the lab processing of samples.
An approach based purely on image classification was proposed in [8]. In this work, the authors proposed using a CNN with an Inception-3 encoder and distinguishing 10–15 lithology classes. They trained the model on images where the dominant lithology is available. Patches of 512 by 512 pixels were fed into the network during training, and the final prediction was obtained by averaging patches results on unseen images.
The lithology identification of mixtures of different lithotypes was considered in [9]. To find the amount of each lithotype on the image, the authors used a segmentation network. Firstly, they trained it using single lithotype images and a constant segmentation mask and, after that, applied the trained network to the images of mixtures and computed the proportion of each lithotype on the photo. The authors noted the main challenge in the lithology detection of drill cuttings, namely very high intra-class variations and low inter-class variations. To overcome this difficulty, the authors split the classes into several subclasses based on the visual appearance of cuttings—grain size, color, etc.—resulting in a total class number of 72 instead of 3 initial rock types: carbonate, sandstone, and shale. The authors did not mention whether subclass annotation was performed manually or automatically. The authors mentioned the work of [10], where it was shown that it is possible to learn more universal representations using a sub-classes approach. The authors applied K-means clustering to image embeddings and used cluster labels as additional features to initial class labels for training a new model.
The work in [11] continued the idea of learning more universal representations by applying clustering to embeddings. The authors used different numbers of clusters, from 500 to 15,000 (having approximately 1000 classes in the initial dataset labeling). It was proposed to add the cluster labels as additional labels to train the new model on this dataset.
In this work, we tried to improve drill cuttings classification using dataset-cleaning techniques, clustering for identifying possibly erroneous classifications, and oversampling for a better separation of visually similar images.

2. Materials and Methods

Within this section, we (1) develop a workflow for lithology prediction from drill cuttings images, (2) describe a suggested approach to eliminate samples with errors in labels, (3) demonstrate the architecture of the applied neural network, and (4) describe the approach used to assess the prediction abilities of the neural network.

2.1. General Workflow for Lithology Prediction from Drill Cuttings’ Images

The developed workflow for predicting rock lithology from drill cuttings images is depicted in Figure 1 and includes the following principal steps:
  • Data gathering. The data include drill cuttings images and their labels. For labels, we imply lithology class provided by the field geologist. Cuttings should be taken under the same lightning conditions and fixed camera setup (shutter speed, aperture, sensitivity (gain), etc.). The samples of drill cuttings should be preliminary washed and dried to be informative with respect to their geological content.
  • Data quality control. Due to human errors, the data always contain some portion of samples with a poor quality of images or their labels. For poor-quality images, we assume images with mud contamination and images with bad exposure. To eliminate the images with mud contamination, a CNN with a binary output is used. Images with bad exposure can be detected by simple image processing techniques (e.g., assigning some threshold values for the luminosity). Errors during labeling is not a rare case due to (1) high variability of visual appearance within the same lithotype and (2) similarity between certain lithologies. The implemented approach for detecting samples with errors in sample labels is presented in Section 2.2.
  • Data preparation. This includes (1) generation of pseudo-labels and (2) data balancing. The generation of pseudo-labels is motivated by diversity of visual appearance within each lithotype. To generate pseudo-labels, we suggest utilizing clusterization of images based on their representations (embeddings). As in many other classification problems, data balancing is a crucial aspect to consider before model training. Depending on the geological profile type, some lithotypes are over-represented and others are represented in a limited quantity. The suggested balancing strategy is presented in Section 2.3.
  • Model training and validation. After the data are prepared, the model training and validation are conducted. To imitate the real case scenario, we subdivided the data into three subsets: (1) training subset, (2) validation subset, and (3) testing subset with unseen data from one well. The training and validation subsets are sampled randomly from the dataset (excluding testing subset) in the ratio of 80% and 20%, respectively. The model architecture is described in subsection “CNN model architecture”.
  • Lithology prediction from digital images of drill cuttings. Once the model is trained, it can be utilized for predicting rock lithology from digital images of drill cuttings. The results of predicting data for test well are described within the Section 3.

2.2. Detecting Samples with Wrong Labels

To identify the samples with errors in labeling, we relied on the following assumption. The dataset contains clusters with one dominant lithotype, and a minor portion of the data within each cluster are presented by samples with wrong labels. Thus, we conducted the quality control procedure in the following way:
  • Obtaining image embeddings. During our research, we used image embeddings in the form of fully connected hidden layer of the CNN.
  • Clustering image embeddings by means of Gaussian mixture algorithm (see, e.g., [12] by Ouyang et al.). To understand the optimal number of number of mixtures, we assessed the silhouette score [13], taking the maximum value of it.
  • Analyzing and processing the data within the detected clusters. During this step, we analyzed the frequency of the lithotype within the cluster. The samples with a low frequency were eliminated from the dataset.
Additionally, the samples should be sanitized for mud contamination and bad exposition during photographing of the samples. The mud-contaminated images can be identified by application of trained CNN that has a binary output (0 for clean sample and 1 for contaminated sample). The bad exposition of images can be determined by simple image processing techniques, e.g., we can calculate the percentage of images that have high values in RGB channels.

2.3. Data Balancing

The data balancing process schema is presented in Figure 2 and consisted of several principal stages (see [14]):
  • Selecting overrepresented lithotypes and removing from consideration a set of images with such lithotypes.
  • Finding unrepresented lithotypes (or present in negligibly small amount) and removing them from the dataset.
  • Determining the number to repeat (one or another sample) in the set so that, as a result, the total number of samples within each lithotype is approximately the same. We denote the number of repetitions by a vector x, where the length of x is equal to the number of remaining samples, and the x components are non-negative integers equal to the number of repetitions of a particular sample. From the label vector of remaining samples, we form a matrix M, where the columns are lithology vectors. The vector x can be found by minimizing the difference between quantities of different lithotypes:
M x M x L 2   x min
It is worth noting that, when dealing with an imbalanced dataset, incorporating class weights during the training process can be useful. However, this may lead to a situation where an under-represented class rarely appears in the training batch for the network, and thus the network infrequently ‘encounters’ such a class during training. By repeating images of an under-represented class, it will appear in the data batches with the same frequency as the others, and additional image augmentation will aid in diversifying the visual representation of this class.

2.4. CNN Model Architecture

After testing distinct architectures of neural networks and analyzing the models’ performance and the resources required for model training, we selected the following architecture of the model (see Figure 3).
We selected MobileNet2 encoder, since it reportedly has good performance. Following [15], where the authors reported that “higher-capacity models perform better on the subset of incorrectly labeled test data in terms of their accuracy on the original labels”, we chose the architecture to be a compromise between size and performance, since our labels presumably contain errors, which are quite difficult to notice.

2.5. Assessment of Prediction Results

From comparison of true and predicted values of rock lithology, we calculated accuracy (Equation (1)), recall (Equation (2)), precision (Equation (3)), and F1 score (Equation (4)) for each lithotype separately. Additionally, we assessed macro and weighted average values for precision, recall, and F1 score.
A c c u r a c y = T P T P + F P + T N + F N
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
F 1 = 2 ·   P r e c i s i o n   · R e c a l l P r e c i s i o n + R e c a l l
where TP is true positive predicted values, FP is false positive predicted values, TN is true negative predicted values, and FN is false negative predicted values.

2.6. Dataset Overview

Within our research, we gathered 16,687 cuttings images from 28 industrial wells in 22 distinct hydrocarbon fields that were captured under white light with a fixed camera set-up. The dataset contains the following lithotypes: (1) anhydrite (ANHY), argillaceous limestone (ARGLS), calcareous limestone (CALLS), clay, dolomite (DOL), dolomitic limestone (DOLLS), halite (HAL), limestone (LS), marl, sandstone (SAND), shale (SH), and siltstone (SST). Figure 4 illustrates several examples of drill cuttings images and a microscope used for making images.
Although external lighting conditions may slightly change during the photography process, artificial light sources were employed to minimize these variations. The camera parameters were manually set and remained constant, ensuring that shutter speed, aperture, and color representation in the images were not significantly altered. The depth from which each sample of drill cutting was retrieved was encoded in the file name.
The gathered dataset exhibits a high degree of imbalance with respect to the total number of samples within distinct lithotypes (Figure 5). That imbalance is common for the study region. As can be seen from Figure 5, the dominant lithology is limestone, whereas the total number of samples for marl, clay, calcareous limestone, halite, siltstone, and dolomite is less than 1000.
As was mentioned, the data come from distinct wells drilled through 101 geological formations occurring at different depth intervals. Due to different sedimentation environment within specific formations, there is a diversity of visual appearance of the same lithotypes coming from distinct formations. Figure 6 shows an example of siltstone samples coming from distinct depth intervals and formations. This problem conditions the necessity to generate pseudo-labels within each class to account for the domain specifics.

3. Results and Discussion

To evaluate the effectiveness of the proposed workflow, we conducted a set of experiments with different approaches to data preparation and model training. For testing the obtained models, we used data from an unseen well that was not used during training or validation. These experiments were:
  • Baseline model. The dataset is taken as is and the model output contains only original lithology labels.
  • Model training on sanitized dataset (removing mud-contaminated images, images with bad exposition, and samples with errors detected my means of the method described in Section 3.2).
  • Model training on sanitized dataset and generated sub-labels by means of data clustering.
During model training, we used the adaptive moment estimation algorithm with a fixed learning rate of 0.0001. During model training, we set the batch size (number of samples processed before the model is updated) to 50 samples.

3.1. Training Baseline Model

During the first experiment, the model for lithology prediction from cuttings images was trained without any data quality control and preprocessing. For training and validation, we took all images and randomly split them into train and validation subsets from the same wells. We applied standard augmentation techniques on the images, including rotation, shift, and skew with little color modification. The training history is depicted in the left panel of Figure 7, whereas the confusion matrix for the validation subset is presented in the right panel.
The obtained results allow us to conclude that the model converged well, with increasing accuracy curves (Figure 7). The prediction accuracy for validation images in each well ranged between 70% and 95%. As depicted in Figure 7, the model demonstrates strong generalization capabilities when both training and validation images are from the same set of wells.

3.2. Results of Dataset Quality Control

To detect images with bad exposure, we specified threshold values for the percentage of pixels that are ‘too white’ or ‘too black’. For ‘too white’ images, we calculated the percentage of pixels that have minimum values along each RGB channel greater than 254. The threshold value of the percentage was set to 3.5% for considering the image as overexposed. For ‘too black’ images, we calculated the percentage of pixels that have maximum values along each RGB channel less than 10. The threshold value of the percentage was set to 10% for considering the image as underexposed. These threshold values were selected iteratively controlling the visual appearance of the images.
To identify images of rocks covered with mud (examples illustrated in Figure 8), we first manually selected a subset of such contaminated images and trained a binary classifier that predicted mud contamination. After that, this model was applied to the whole dataset and revealed images with contamination. The total percentage of images detected as contaminated was 5% from the overall dataset, including the test well.
After that, we detected samples with problems with labeling. First, we applied a cross-validation technique: we divided all data into ten folds and performed training ten times, each time using one of the folds for model evaluation and the rest of the images for training. The uncertainly labeled images were those with a maximal error (when the model predicted a 100% single-lithotype image with the wrong lithotype) and highest predicted confidence. Figure 9 shows the plot of root mean square error (RMSE) versus prediction confidence for one of the runs. (Note that, in case a 100% single lithotype was misclassified, RMSE = 100 2 .) The uncertain images for all runs of cross-validation were marked as having uncertain labels.
The model trained on sanitized data, after cross-validation, was used for image evaluation. To estimate how reliable the classification result can be, we applied clustering of the vectors embeddings of train images produced by this model. We applied the K-means clustering algorithm, where the number of clusters was estimated using the silhouette score after several clustering runs. The following Figure 10 illustrates the distribution of lithotypes in different clusters, where the number of clusters is equal to 150. The bar plot in Figure 10 shows the counts of samples in each cluster, and the color of the bars is attributed to different lithotypes. From this bar plot, we can see that each cluster can contain different lithotypes and a different count of samples of the lithotypes.
The clusters contain varying quantities of lithotypes, while images within each cluster exhibit visual similarities. Usually, visual similarity between two images can be estimated as the Euclidean distance between vectors embeddings. Consequently, we can assess the reliability of labels for each cluster. To accomplish this, we identified the lithotype corresponding to the maximum and second maximum number of images within a specific cluster. If the maximum number is substantially larger than the second maximum, we have a relatively uniform cluster, indicating that visually similar images are labeled with the same lithotype. To attribute a cluster to a set of ‘good’ or ‘bad’ clusters, we used a threshold of 0.1, meaning that if the maximum is 10 times greater than the second maximum, we count the cluster as ‘good’; otherwise, it is ’bad’ (non-reliable). The following figures (Figure 11, Figure 12 and Figure 13) illustrate several clusters: on the top row, images from the train set are given; on the bottom row, images from the test set with the predicted lithotype using the dominant lithotype of the particular cluster, where the true label is in brackets, are given.
The prediction accuracy that we obtained using the clustering method is presented in Table 1 below.
We can see that the prediction accuracy significantly drops in clusters where several lithotypes are present. In this case, we can ask for a re-labeling of these images or drop them from the training set.

3.3. Generation of Sublabels for the Dataset

Another approach that we applied is based on creating label sub-classes. To account for diversity within each lithotype, we relabeled our samples based on clustering in the following way. Firstly, we obtained embeddings for each sample by means of the baseline model. The shape of embeddings was 1280 × 1. Secondly, we conducted a dimensionality reduction of these embeddings by means of principal component analysis implemented within the scikit-learn library [16] to the shape 100 × 1. In the third step, we clustered the obtained embeddings within each lithotype separately using the Gaussian mixture algorithm. To detect the optimal number of clusters, we specified the range constraining the maximum of possible clusters to the total number of geological formations where the lithotype occurred. The optimal value of clusters was determined by the maximum value of the above-mentioned silhouette score. The results of data clustering are summarized within Table 2. The total number of samples within each cluster is depicted in Figure 14.

3.4. Data Balancing and Model Training

After the data quality control and generation of sublabels, data balancing was carried out following the general workflow (see Figure 2). The applied algorithm for data balancing is described within Section 2.3. Figure 15 illustrates the results of data balancing for the dataset without the additional generation of sublabels. Figure 16 depicts the results of data balancing for the dataset with additionally generated sublabels.
Model training history for data without and with the generation of sublabels for lithotypes is presented in Figure 17. As can be seen from Figure 17, the generation of sublabels resulted in a faster convergence of the prediction accuracy for the validation subset.

3.5. Predicting Rock Lithology for Samples from Unseen Well

The three trained models (the baseline model, the model trained on the sanitized dataset, and the model trained on the sanitized dataset with the generation of sublabels within lithotypes) were applied to predict rock lithology from cuttings images of the unseen well. From the comparison of the true and predicted classes of lithotypes, we assessed the performance of models (Table 3).
From the obtained results, we can conclude that, when the baseline model is applied to images from a different well not used in training and validation, the model’s performance deteriorates significantly (28% accuracy). In this case, the lithology estimation is rather coarse, with some lithotypes predicted more inaccurately than others. This could be attributed to varying conditions for sample preparation, rock grinding, and drilling fluid composition, and a significant diversity in the training set. From these results, we can see that the whole dataset contains a lot of uncertainties, which limit the models’ performance.
At the same time, the conducted quality control of the dataset resulted in a significant enhancement of the model quality (the accuracy increased to 51). This evidences that quality control is one of the key aspects enhancing the generalization ability of the model.
The generation of sublabels allowed for accounting for diversity within each lithotypes and resulted in an additional increase in prediction metrics (to a 60% accuracy). The obtained results correlate with research findings presented in [10,11].

4. Conclusions

A new approach for predicting rock lithology from cuttings images was developed and tested on experimental data from 28 industrial wells. The developed approach relies on the application of a convolutional neural network. The utilized CNN architecture represents an adaptation of the MobileNet2 for the problem. A special data preparation workflow was implemented within the experiment to account for (1) the diversity of visual appearance within each lithological class and (2) errors in image labels. The diversity of the visual appearance was accounted for through clustering each lithotype separately and generating new sub-classes based on the clustering results. The errors in image labeling were detected based on the results of clustering the whole dataset and analyzing the classes distribution within the obtained clusters. Within the case study, we trained three models: (1) a baseline model, (2) a model trained on a sanitized dataset, and (3) a model trained on a dataset with new sublabels generated by means of clustering. The performance of the models was assessed on data from an unseen well. The obtained results reveal a significant increase in prediction metrics due to the data preparation step. This evidences that the data preparation step plays one of the key roles during the enhancement of the generalization ability of the model for predicting cuttings lithology. The model trained on generated sublabels via clustering also provided a prominent increase in prediction metrics for the unseen well. The obtained results show that, in the best case, the F1 score on the unseen well can reach 66%, which is an acceptable level of prediction uncertainty. The implementation of the developed workflow can sufficiently reduce the time spent by geologists on cuttings description and can eliminate human errors. The next step for the research direction could be testing and enhancing the proposed solution for samples with mixed lithology. This problem can be classified as regression-type and will obviously require additional efforts.

Author Contributions

Conceptualization, investigation, software, E.T., A.S. and M.M.; resources, data curation, project administration, M.M. and S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are unavailable due to privacy restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.; Yang, S.; Zhao, Y.; Wang, Y. Lithology identification using an optimized KNN clustering method based on entropy-weighed cosine distance in Mesozoic strata of Gaoqing field, Jiyang depression. J. Pet. Sci. Eng. 2018, 166, 157–174. [Google Scholar] [CrossRef]
  2. Sun, J.; Chen, M.; Li, Q.; Ren, L.; Dou, M.; Zhang, J. A new method for predicting formation lithology while drilling at horizontal well bit. J. Pet. Sci. Eng. 2021, 196, 107955. [Google Scholar] [CrossRef]
  3. Li, Z.; Wu, Y.; Kang, Y.; Lv, W.; Feng, D.; Yuan, C. Feature-Depth Smoothness Based Semi-Supervised Weighted Extreme Learning Machine for lithology identification. J. Nat. Gas Sci. Eng. 2021, 96, 104306. [Google Scholar] [CrossRef]
  4. Chen, G.; Chen, M.; Hong, G.; Lu, Y.; Zhou, B.; Gao, Y. A new method of lithology classification based on convolu-tional neural network algorithm by utilizing drilling string vibration data. Energies 2020, 13, 888. [Google Scholar] [CrossRef] [Green Version]
  5. Kathrada, M.; Adillah, B.J. Visual Recognition of Drill Cuttings Lithologies Using Convolutional Neural Networks to Aid Reservoir Characterisation. In Proceedings of the SPE Reservoir Characterisation and Simulation Conference and Exhibition, Abu Dhabi, United Arab Emirates, 17–19 September 2019. [Google Scholar] [CrossRef]
  6. Huo, F.; Li, A.; Zhao, X.; Ren, W.; Dong, H.; Yang, J. Novel lithology identification method for drilling cuttings under PDC bit condition. J. Pet. Sci. Eng. 2021, 205, 108898. [Google Scholar] [CrossRef]
  7. Xu, Z.; Shi, H.; Lin, P.; Liu, T. Integrated lithology identification based on images and elemental data from rocks. J. Pet. Sci. Eng. 2021, 205, 108853. [Google Scholar] [CrossRef]
  8. Equinor. Cuillin. Available online: https://www.npd.no/globalassets/2-force/2019/documents/archive-2010-2018/force-hackathon/nr.15-wade-cuillin.pdf/ (accessed on 10 May 2023).
  9. Tamaazousti, Y.; François, M.; Kherroubi, J. Automated identification and quantification of rock types from drill cuttings. In SEG Technical Program Expanded Abstracts; Society of Exploration Geophysicists: Houston, TX, USA, 2020; pp. 1591–1595. [Google Scholar] [CrossRef]
  10. Girard, J.; Tamaazousti, Y.; Borgne, H.L.; Hudelot, C. Learning finer-class networks for universal representations. arXiv 2018, arXiv:1810.02126. [Google Scholar]
  11. Yan, X.; Misra, I.; Gupta, A.; Ghadiyaram, D.; Mahajan, D. Clusterfit: Improving generalization of visual represen-tations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6509–6518. [Google Scholar]
  12. Ouyang, M.; Welsh, W.J.; Georgopoulos, P. Gaussian mixture clustering and imputation of microarray data. Bioinformatics 2004, 20, 917–923. [Google Scholar] [CrossRef] [PubMed]
  13. Shahapure, K.R.; Nicholas, C. Cluster quality analysis using silhouette score. In Proceedings of the 2020 IEEE 7th International Conference on Data Science and Advanced Analytics (DSAA), Sydney, Australia, 6–9 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 747–748. [Google Scholar]
  14. Mezghani, M.; Tolstaya, E. Toward Autonomous Wellsite Geology: Artificial Intelligence for Cutting Li-thology Prediction. In Proceedings of the Second EAGE Subsurface Intelligence Workshop, Manama, Bahrain, 28–31 October 2022; EAGE Publications BV: Houten, The Netherlands, 2022; Volume 2022, pp. 1–5. [Google Scholar]
  15. Northcutt, C.G.; Athalye, A.; Mueller, J. Pervasive label errors in test sets destabilize machine learning benchmarks. arXiv 2021, arXiv:2103.14749. [Google Scholar]
  16. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
Figure 1. Workflow for predicting rock lithology from drill cuttings images by means of deep learning.
Figure 1. Workflow for predicting rock lithology from drill cuttings images by means of deep learning.
Jimaging 09 00126 g001
Figure 2. Dataset balancing.
Figure 2. Dataset balancing.
Jimaging 09 00126 g002
Figure 3. Deep learning model architecture. N_lith stands for total number of lithotypes to predict.
Figure 3. Deep learning model architecture. N_lith stands for total number of lithotypes to predict.
Jimaging 09 00126 g003
Figure 4. Examples of drill cutting images for different lithotypes.
Figure 4. Examples of drill cutting images for different lithotypes.
Jimaging 09 00126 g004
Figure 5. Bar plot representing total number of samples per lithotype within the dataset.
Figure 5. Bar plot representing total number of samples per lithotype within the dataset.
Jimaging 09 00126 g005
Figure 6. Example of distinct visual appearance for siltstone coming from different depth intervals and formations.
Figure 6. Example of distinct visual appearance for siltstone coming from different depth intervals and formations.
Jimaging 09 00126 g006
Figure 7. Accuracy curves for classifier training on 100% single lithotypes images (left panel); confusion matrix (right panel).
Figure 7. Accuracy curves for classifier training on 100% single lithotypes images (left panel); confusion matrix (right panel).
Jimaging 09 00126 g007
Figure 8. Examples of bad images with mud contamination.
Figure 8. Examples of bad images with mud contamination.
Jimaging 09 00126 g008
Figure 9. Plot of RMSE error versus prediction confidence for one of the cross-validation runs.
Figure 9. Plot of RMSE error versus prediction confidence for one of the cross-validation runs.
Jimaging 09 00126 g009
Figure 10. Distribution of lithotypes in clusters.
Figure 10. Distribution of lithotypes in clusters.
Jimaging 09 00126 g010
Figure 11. Cluster with 72 samples of argillaceous limestone and 17 samples of limestone (top row, from train set; bottom row, from test set; true label in brackets). Here, visually similar images have different labels; therefore, labeling is not reliable.
Figure 11. Cluster with 72 samples of argillaceous limestone and 17 samples of limestone (top row, from train set; bottom row, from test set; true label in brackets). Here, visually similar images have different labels; therefore, labeling is not reliable.
Jimaging 09 00126 g011
Figure 12. Cluster with 86 samples of dolomitic limestone and 83 samples of limestone (top row, from train set; bottom row, from test set; true label in brackets). Here, visually similar images have different labels; therefore, labeling is not reliable.
Figure 12. Cluster with 86 samples of dolomitic limestone and 83 samples of limestone (top row, from train set; bottom row, from test set; true label in brackets). Here, visually similar images have different labels; therefore, labeling is not reliable.
Jimaging 09 00126 g012
Figure 13. Cluster where the amount of most represented lithotype is larger (in our experiment, 10 times larger) than second-most represented lithotype, providing good classification results (top row, from train set; bottom row, from test set; true label in brackets).
Figure 13. Cluster where the amount of most represented lithotype is larger (in our experiment, 10 times larger) than second-most represented lithotype, providing good classification results (top row, from train set; bottom row, from test set; true label in brackets).
Jimaging 09 00126 g013
Figure 14. Bar plot representing total number of samples per sub-cluster of lithotypes.
Figure 14. Bar plot representing total number of samples per sub-cluster of lithotypes.
Jimaging 09 00126 g014
Figure 15. Results of data balancing for original dataset.
Figure 15. Results of data balancing for original dataset.
Jimaging 09 00126 g015
Figure 16. Results of data balancing for dataset with sublabels. Numbers after lithology short names indicate the relevance to the clusters.
Figure 16. Results of data balancing for dataset with sublabels. Numbers after lithology short names indicate the relevance to the clusters.
Jimaging 09 00126 g016
Figure 17. Accuracy curves for classifier training on the dataset without (A) and with (B) additional sublabels after data balancing.
Figure 17. Accuracy curves for classifier training on the dataset without (A) and with (B) additional sublabels after data balancing.
Jimaging 09 00126 g017
Table 1. Lithology prediction accuracy using clustering.
Table 1. Lithology prediction accuracy using clustering.
TrainTestSamples Number
‘Good’ clusters98%51%5451
‘Bad’ clusters70%28%3476
Table 2. Results of data clustering within each lithotype.
Table 2. Results of data clustering within each lithotype.
LithotypeN of FormationsOptimal N of Clusters
Anhydrite53
Argillaceous limestone107
Calcareous limestone64
Clay54
Dolomite95
Dolomitic limestone1814
Halite63
Limestone1714
Marl11
Sandstone2016
Shale118
Siltstone1510
Table 3. Results of predicting rock lithology for cuttings images from unseen well.
Table 3. Results of predicting rock lithology for cuttings images from unseen well.
LithotypePrecisionRecallF1Accuracy
Macro AvrWeighted AvrMacro AvrWeighted AvrMacro AvrWeighted Avr
Baseline model0.260.780.290.460.230.560.28
Sanitized data model0.320.710.400.50.300.610.42
Sub-labels model0.360.770.450.60.340.660.60
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tolstaya, E.; Shakirov, A.; Mezghani, M.; Safonov, S. Fast Reservoir Characterization with AI-Based Lithology Prediction Using Drill Cuttings Images and Noisy Labels. J. Imaging 2023, 9, 126. https://doi.org/10.3390/jimaging9070126

AMA Style

Tolstaya E, Shakirov A, Mezghani M, Safonov S. Fast Reservoir Characterization with AI-Based Lithology Prediction Using Drill Cuttings Images and Noisy Labels. Journal of Imaging. 2023; 9(7):126. https://doi.org/10.3390/jimaging9070126

Chicago/Turabian Style

Tolstaya, Ekaterina, Anuar Shakirov, Mokhles Mezghani, and Sergey Safonov. 2023. "Fast Reservoir Characterization with AI-Based Lithology Prediction Using Drill Cuttings Images and Noisy Labels" Journal of Imaging 9, no. 7: 126. https://doi.org/10.3390/jimaging9070126

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop