Next Article in Journal
EP-ADTA: Edge Prediction-Based Adaptive Data Transfer Algorithm for Underwater Wireless Sensor Networks (UWSNs)
Next Article in Special Issue
Image Recovery from Synthetic Noise Artifacts in CT Scans Using Modified U-Net
Previous Article in Journal
Learning-Based Slip Detection for Robotic Fruit Grasping and Manipulation under Leaf Interference
Previous Article in Special Issue
Multi-Scale Attention Convolutional Network for Masson Stained Bile Duct Segmentation from Liver Pathology Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Precancerous Lesion Screening Using an Instance Segmentation Technique for Improving Accuracy

by
Patiyus Agustiansyah
1,2,
Siti Nurmaini
3,*,
Laila Nuranna
4,
Irfannuddin Irfannuddin
5,
Rizal Sanif
5,
Legiran Legiran
5,
Muhammad Naufal Rachmatullah
3,
Gavira Olipa Florina
3,
Ade Iriani Sapitri
3 and
Annisa Darmawahyuni
3
1
Doctoral Program, Biology Science, Faculty of Medicine, Universitas Sriwijaya, Palembang 30139, Indonesia
2
Division of Oncology-Gynecology, Department of Obstetrics and Gynecology, Mohammad Hoesin General Hospital, Palembang 30126, Indonesia
3
Intelligent System Research Group, Faculty of Computer Science, Universitas Sriwijaya, Palembang 30139, Indonesia
4
Obstetrics & Gynecology Department, Faculty of Medicine, University of Indonesia, Jakarta 10430, Indonesia
5
Obstetrics & Gynecology Department, Faculty of Medicine, Universitas Sriwijaya, Palembang 30139, Indonesia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(15), 5489; https://doi.org/10.3390/s22155489
Submission received: 30 May 2022 / Revised: 13 July 2022 / Accepted: 17 July 2022 / Published: 22 July 2022
(This article belongs to the Special Issue Artificial Intelligence-Based Applications in Medical Imaging)

Abstract

:
Precancerous screening using visual inspection with acetic acid (VIA) is suggested by the World Health Organization (WHO) for low–middle-income countries (LMICs). However, because of the limited number of gynecological oncologist clinicians in LMICs, VIA screening is primarily performed by general clinicians, nurses, or midwives (called medical workers). However, not being able to recognize the significant pathophysiology of human papilloma virus (HPV) infection in terms of the columnar epithelial-cell, squamous epithelial-cell, and white-spot regions with abnormal blood vessels may be further aggravated by VIA screening, which achieves a wide range of sensitivity (49–98%) and specificity (75–91%); this might lead to a false result and high interobserver variances. Hence, the automated detection of the columnar area (CA), subepithelial region of the squamocolumnar junction (SCJ), and acetowhite (AW) lesions is needed to support an accurate diagnosis. This study proposes a mask-RCNN architecture to simultaneously segment, classify, and detect CA and AW lesions. We conducted several experiments using 262 images of VIA+ cervicograms, and 222 images of VIA−cervicograms. The proposed model provided a satisfactory intersection over union performance for the CA of about 63.60%, and AW lesions of about 73.98%. The dice similarity coefficient performance was about 75.67% for the CA and about 80.49% for the AW lesion. It also performed well in cervical-cancer precursor-lesion detection, with a mean average precision of about 86.90% for the CA and of about 100% for the AW lesion, while also achieving 100% sensitivity and 92% specificity. Our proposed model with the instance segmentation approach can segment, detect, and classify cervical-cancer precursor lesions with satisfying performance only from a VIA cervicogram.

1. Introduction

Cervical cancer has become a public health concern and global health burden, with approximately 90% of fatalities coming from low–middle-income countries (LMICs) [1]. Following a 27% increase in fatality rate by 2030, the frequency of cervical cancer is expected to increase to 21% [1,2]. Increasing the accurate and thorough coverage screening of cervical cancer precursor lesions in LMICs is a main method for reducing the incidence and fatality rate of the disease [3,4,5,6]. Cervical cancer is a preventable disease caused by an oncogenic form of the human papilloma virus (HPV), which is an oncogenic variant/high-risk type [7]. According to Herf’s pathophysiology of HPV infection [4,5], viral infections originate from the subepithelial region of the squamocolumnar junction (SCJ) and spread into the cervical transformation zone (TJ). To diagnose this condition, clinicians must be able to recognize the columnar area (CA), SCJ subepithelial regions, and acetowhite (AW) lesions. If the subepithelial SCJ region and AW lesion do not overlap, they are defined as condyloma lesions (which are created by low-risk HPV virus) or Nabothian cysts (blocked Nabothian gland). The pathophysiology of HPV infection is a crucial anatomical characteristic for diagnosing cervical-cancer precursor lesions [8,9,10].
Cervical-cancer precursor lesions can progress to invasive carcinoma after about 10–15 years [11,12,13,14]; hence, there is a “golden time” to perform early screening to break the disease’s chain. Currently, precancerous lesion screening consists of various methods, including a Papanicolaou (Pap) smear based on liquid-base cytology (LBC), visual inspection with acetic acid (VIA), visual inspection with Lugol’s iodine (VILI), and HPV DNA test based on genotyping or hybrid capture [14]. Pap smear screening programs have been successful in reducing the burden of cervical cancer in the developed world [3,8]. In LMICs, it is not a feasible option because of the lack of trained cytopathologists, difficulties in following up with screen-positive women, the absence of quality assurance measures, and poor healthcare infrastructures [1,2,8]. VIA screening is recommended by the WHO for LMICs because of its low cost and simple process, and it can be performed in a real-time screening test. In every VIA+ case, cold ablation (cryotherapy) intervention is conducted at the same visit to prevent the loss of patient follow-up [6]. However, VIA has the limitations of low specificity, requiring extensive training, and retraining healthcare providers, especially with the test being subjective and challenging regarding quality control [15,16,17,18,19,20]. Hence, automated VIA screening that can achieve accurate and satisfactory performance is needed.
A diagnostic medical procedure to take pictures of the cervix for interpretation on the basis of VIA produces a cervical digital image named a cervicogram. The cervicogram provides the permanent and objective documentation of normal and abnormal anatomical cervical patterns [4,10]. The automated analysis of these important patterns is a challenging task because of the following [5,7,16]: (i) the presence of artifacts such as shadows in the cervical area during the cervicogram acquisition process, which can be caused by light reflections and can produce a concave shape of the cervix; (ii) the complex variability in cervicogram image data in terms of intensity, shape, the presence of ovules of Nabothi, and areas of immature metaplasia; (iii) variability in the cervicogram content because not all cervical tissue appears in the cervicogram; and (iv) the presence of a narrow dynamic color threshold and the unclear boundaries of the tissue region.
Computer-aided diagnosis (CAD) studies based on an artificial intelligence (AI) approach have produced remarkable results in the medical field [17,18,19,20,21,22,23,24,25]. These technologies can automatically predict abnormalities to support patients’ diagnoses and offer medical personnel a second opinion. CAD-based AI methods include machine learning (ML) and deep learning (DL) [21,22,23,24,25]. They can generate a learning model to enhance diagnostic accuracy in a wide variety of clinical sectors. In precancerous detection, various approaches for analyzing cervicogram images have been investigated, including the use of handicraft characteristics and conventional classification algorithms such as support vector machines (SVMs) [17] and K-nearest neighbor (K-NN) [18]. However, ML cannot automatically conduct feature extraction and still requires human intervention. This limits the generalization model in routine clinical applications because of the significant variability in the pathophysiology of the cervicogram images.
The DL model has been proposed to identify precursor cervical-cancer lesions by segmenting, detecting, or classifying them using feature learning. Compared with ML, deep-learning algorithms possess more powerful learning capabilities and can automatically extract features without extensive data preprocessing or handcrafted feature extraction, rendering them a suitable tool for analyzing the complex structures of high-dimensional data [24]. Several studies have been developed to aid diagnoses here by using cervicography and colposcopy images, which are explored in VIA screening to detect AW lesions as symptoms of precursor cervical cancer. These studies used models employing faster region convolutional neural network (CNN) architectures [21], DL models [22], and CNN architectures [23,24]. All these studies show satisfying results, and our model outperformed other ML models.
However, the processes of segmentation, detection, and classification by using cervicography are separately carried out. It is difficult to analyze in great detail where the position of an AW lesion is because of the classification running in the black-box condition. In addition, the classification of AW lesions requires huge datasets; unfortunately, it is hard to collect VIA+ cervicograms. The success of model evaluation depends on the accurate segmentation of CA and AW lesions with the SCJ. Despite recent improvements in this field, the presence of AW lesions, such as condyloma and nabothian cysts, may cause system misinterpretation and impair accuracy. To improve prediction accuracy, our research focuses on segmenting two key aspects that explain the crossing of CA and AW lesions. The proposed model also detects CA and AW lesions simultaneously. Here, mask region networks were employed using the CNN technique for approaching deep learning to interpret cervicogram images; this also helps in the human side, adding faster learning without vast datasets. The contributions and originality of our study are as follows:
  • An accurate segmentation of the CA and AW lesion is proposed based on the pathogenesis of cervical cancer by HPV with high confidence and a high intersection over union (IoU) baseline.
  • An instance segmentation model was developed with a mask-RCNN architecture to simultaneously segment, classify, and detect AW lesions in VIA+ and VIA−.
  • The proposed model was validated within a real clinical setting to ensure that the model could be trusted in terms of false-positive and false-negative predictive values.
The current paper is structured as follows: In Section 2, we present the research methods proposed for cervical-cancer precursor-lesion detection. In Section 3, we present the results of applying the methods and a discussion of these results. Lastly, in Section 4, we draw conclusions.

2. Materials and Methods

The entire methodology was divided into five main processes: data acquisition, data preparation, annotation labels, DL model training, and model evaluation. The DL model was mainly used to automatically segment, detect, and classify CA and AW lesions. The proposed DL model used in the current study is shown in Figure 1. More details about each step are given in the next subsection.

2.1. Data Preparation

In the present study, two experienced senior gynecological oncologists defined the important anatomic landmarks to label the cervicogram images. The cervicograms were taken from Mohammad Hoesin Indonesian General Hospital. We performed four steps of preprocessing: (i) collecting digital cervical images; (ii) selecting cervicogram images with adequate pathophysiology and clear AW lesions; (iii) removing unnecessary information with the image cropping process; and (iv) resizing the cervicogram images to produce the same size for all images. The cervicograms were taken using a mobile camera after the application of acetic acid to the cervical region. Due to the utilization of different mobile cameras to capture the cervicograms, the dimensions of the images varied from 150 × 130 pixels to 1024 × 1027 pixels. All cervicograms should have adequate landmarks, such as the CA, subepithelial SCJ region, and AW lesions. Figure 2a shows that red in the endocervix, which is composed of one layer of the columnar cell, appears as the red region, which differs from the TZ, here consisting of multiple layers of squamous cells, which appear as the pink region (Figure 2b). The border between these two different areas is the SCJ. The AW lesion appears as the white area that rises from the SCJ to the TZ (Figure 2c); however, condyloma lesions, closed nabothian glands, and immature metaplasia cells are shaped almost the same as AW lesions but do not coincide with the SCJ.
The dataset was acquired from 484 patients during standard clinical practice in 2020 and 2021 at Mohammad Hoesin General Hospital, Indonesia. A total of 262 patients provided abnormal VIA+ cervicogram images, and 222 patients provided normal VIA− cervicogram images. A whole cervicogram image has a size of about 512 × 512 pixels. The data distribution from the collected cases was split randomly into training, validation, and testing sets for the learning process, as represented in Table 1. Two gynecological oncologist clinicians selected the most widely used landmarks in routine VIA screening. The gynecological oncologists annotated only images complying with the minimal quality requirements, and only a clear cross-sectional scan image was included for further processing. We conducted the learning process without augmentation data because we wanted to maintain the actual clinical conditions.

2.2. Image Annotation

In the segmentation process, the anatomy landmark of the cervix plays an important role, especially in SCJ, CA, TZ, and AW lesions. Two gynecological oncologists who have over 10 years of experience following the above protocol manually annotated such landmarks using an annotation tool (LabelMe) as the ground truth [25,26]. Both normal (without AW lesion) and abnormal (with AW lesion) cervicogram samples had significant variations in image quality, shape, size, and orientation (Figure 3). Annotated images show the CA and AW lesions in the TZ that intersect with the SCJ (Figure 4). Therefore, it is essential to first recognize the SCJ in every image. The whole ground-truth database was saved in JSON file format.

2.3. Deep Learning Model

The instance segmentation approach was developed in our previous study and it is based on the mask-RCNN architecture here [25,26]. This architecture has two main parts: the region proposal networks (RPNs) that are used as feature extraction processes from raw data, and fully convolutional networks (FCNs) that function as a multitask learning process for simultaneous classification, detection, and segmentation (Figure 5). The mask-RCNN structure allows for the system to generate the class, location (boundary box), and shape of the object of interest [25,26]. The proposed model utilizes a multitask loss function that combines the loss of classification, localization, and mask segmentation, as illustrated in Equation (1).
L T o t a l = L cls + L bbox + L mask
The first term is L cls , which measures the error between the ground-truth and predicted class labels. L cls is the log loss function over the two classes that can be translated from multiclass classification into a binary classification by determining whether the predicted sample is the desired target. p i * is the ground-truth label of anchor box I, and p i is the predicted probability of each anchor i being an object. The log loss function gives L cls (Equation (2)).
L cls   p i , p i * = p i * log p i 1   p i *   log 1 p i
The second loss function is L bbox , which measures the difference between t i u   and v i . t i u   is the predicted result of the bounding box, while v i is the true bounding box. In calculating the loss function   of   L bbox , a smoothing function is added to reduce the sensitivity to outliers. L bbox and the smoothing function are given by Equations (3) and (4):
L bbox = i x , y , w , h L 1 smooth t i u v i
L 1 smooth = 0.5 x 2 , if   x < 1 x 0.5 , else
The third loss is L mask , which calculates the prediction error of the segmentation masks in each instance. The mask branch generates the n × n masking image in every region of interest (RoI) and every class K of the total classes (CA and AW lesions). Hence, the total output was of size K … n2. When the model generated the mask image, there was no competition among the classes because the model learnt from a mask from each class. L mask is the average binary cross-entropy loss only including the k-th mask if the region is associated with ground-truth class k, as defined in Equation (5).
L mask = 1 m 2 1 i , j m   y i j l o g   y ^ i j k +   1 y i j log 1 y ^ i j k
where y i j is the label of a cell (i, j) in the true mask for the region of size m × m, and y ^ i j k is the predicted value of the same cell in the mask learned for ground-truth class k.
For the feature extraction mechanism, the ResNet50 architecture was applied as the backbone in the RPNs to produce a feature map, as shown in Figure 6. All feature maps were generated from different residual blocks. These features provide insights into the internal representation of the specific input to be learned by the FCNs. It also helps in understanding why the model might fail to correctly classify some of the images, hence aiding in finetuning the model for better accuracy and precision.

2.4. Model Evaluation

To validate and evaluate the performance of the instance segmentation model, the outputs of mask-RCNN were validated by using three metrics: the IoU, dice similarity coefficient (DSC) for segmentation, and mean average precision (mAP) for object detection. The IoU is a value based on the statistical similarity and diversity of sample sets; its purpose is to evaluate the overlapping (intersecting) area between two bounding boxes, namely, the predicted and ground-truth bounding boxes. The IoU is formulated in Equation (6).
I o U = a r e a   o f   o v e r l a p a r e a   o f   u n i o n
DSC is a statistical tool that measures the similarity and diversity of sample sets [25,26,27]. In this case, we measured the performance of predictive images with detailed truth labels. The DSC is illustrated in Equation (7).
D S C   X , Y = 2 × a r e a   o f   O v e r l a p T o t a l   N u m b e r   o f   P i x e l s
The mAP score is a widely adopted metric for assessing object-detection models. The mAP values of the various groups were computed, and their averages were obtained. Although the model could detect various objects, the classes assigned to these objects were not always certain. However, even if the expected class for an object or instance is correct, the output criterion must still look at how well the model locates it spatially in the picture. Equation (8) depicts the most commonly used mAP.
mAP = 1 n c l   i = 1 i = n A P i
where ( n c l ) refers to the total of all the different classes, and A P k is the average precision of class i.

3. Results and Discussion

We benchmarked widely used state-of-the-art CNN-based mask-RCNNs with three backbone architectures: ResNet50, ResNet101, and MobileNetV1. The networks’ original architecture of mask-RCNN was maintained in all cervicograms. All networks were first pretrained using the Microsoft Common Objects in Context (COCO) dataset [25,26], and then fully retrained using our training data to produce the probability scores for each class.

3.1. Cervicogram Segmentation and Detection Performan

We conducted the segmentation process using a normal cervicogram with the anatomy of CA, SCJ, and TZ, and an abnormal cervicogram with the anatomy of CA, SCJ, and AW lesions. CA and AW lesions were segmented on the basis of lesion contour, and combined with a bounding box with RoI for lesion detection. The whole images were trained and validated using the mask-RCNN architecture, and three performance metrics were used to assess how well such models work, namely, IoU, DSC, and mAP [28].
In the current study, a mask-RCNN with a ResNet 50 backbone was trained with different learning rates (LR) ranging from 0.001 to 0.00001, with 0.1 increments. Every model was trained using the Adam optimizer, 8 batch size, and 50 epochs. The IoU metric assesses how similar the predicted results are compared with the ground-truth label in the range of 0–1. The greater the IoU value is, the more similar the model prediction with the ground truth label; here, the IoU baseline was set at 0.5. Using a learning rate of about 0.0001, the segmentation process achieved satisfying performance for the IoU, 63.61% for the CA, and 72.43% for the AW lesions (Table 2). In comparison, the DSC assesses how similar the predicted results are to the ground truth by measuring the boundary around the feature. The proposed model achieved a DSC of about 72.55% and 88.81% for the CA and AW lesions, respectively (Table 2).
mAP evaluates how well the model detects a desired object. Detection is accurate if the intersection boundary between the ground truth and prediction has a minimal value of 50% with the same label. From the experiment, the proposed model produced a mAP of about 86.90% for the CA and 100% for AW lesions (Table 2). The unique aspect of our study is that it was validated using Herf’s theory, which states that a true AW lesion always intersects with the SCJ. HPV infection spreads from the subepithelial SCJ that expands to TZ, while a false AW lesion, such as a condyloma lesion, closed nabothian gland, and immature metaplasia cells, do not intersect with the SCJ. The experiment showed that our proposed model can detect a white-spot region with abnormal blood vessels in cervicograms by simultaneously classifying and segmenting the CA.
To analyze the proposed model with the ResNet 50 backbone, we evaluated the model and compared it with other architectures, including ResNet 101 and MobileNetV1. The proposed model with ResNet 50 and 0.0001 LR still produced satisfactory performance compared with the other architectures (Table 3) in terms of the IoU, DSC, and mAP. In addition, AW lesion detection reached 100% in mAP, meaning the model could perfectly detect precancerous lesions.
To find out whether the proposed model could localize objects, recognize each class of objects, and segment lesion contours, training and validation losses were analyzed. There were two losses in the RPNs, three losses in the FCNs, and one average loss (overall loss). Mask-RCNN decouples three tasks in process learning, that is, the bounding-box prediction, the class prediction, and the mask branch, generating mask segmentation for each class without competition among classes. In the present study, the overall loss indicated good performance and converged to one value of about 0.4 in training, and 0.7 in validation (Figure 7).
The classification process was carried out on the basis of two cervicogram images, namely, normal and abnormal anatomical cervical patterns, and the model successfully predicted all normal cervicogram images with 100% sensitivity (Table 4). The classification result based on the ResNet 50 architecture achieved 96.26% accuracy and 92% specificity. Furthermore, the ResNet 50 backbone had the smallest number of false positives and false negatives (Figure 8). A receiver-operating characteristic (ROC) curve was generated to analyze the accuracy of the proposed model. The ROC curve of the ResNet 50 backbone showed the best performance compared with other architectures, with an area under the curve (AUC) value of 0.91 (Figure 8). This means that mask-RCNN could achieve satisfying performance in terms of its segmentation, detection, and classification metrics.
From the experiment, the cervical area in terms of the columnar epithelial-cell (CA), squamous epithelial-cell (SCJ), and white spot (AW) regions with abnormal blood vessels were segmented and precisely detected. As shown in Figure 9, the proposed model achieved higher detection precision in lesion areas with normal and abnormal conditions. The cervix was fully circular, where the SCJ and TZ can be seen in the four screening quadrants (top right, bottom right, top left, and bottom left). The sample prediction images using the proposed model produced AW lesions showing cervical precancerous lesions or VIA+ in Figure 9a–c, whereas only the CA or VIA− is shown in Figure 9d,e. The lesion area with VIA+ or VIA− stratified squamous cell metaplasia was successfully segmented and detected with a confidence value of 90% and IoU over 50%. Cervical intraepithelial neoplasia (CIN) cells have nuclei that are larger and darker than those of normally mature squamous cells. Greasing with 5% acetic acid causes the cytoplasm to shrink and renders the nucleus more prominent. When light hits the CIN, it is reflected back like a mirror, giving it a white appearance. Normal cells are translucent, allowing for light to enter the veins below, giving the cervix its pink appearance. Columnar cells appear to be red because they are one layer above the vascular tissue.

3.2. Benchmarking Our Model with Existing Studies

We benchmarked our proposed DL model with other studies on precancerous prediction. To the best of our knowledge, very limited studies have used DL models based on VIA cervicograms. For a fair comparison, the selected model was benchmarked with the same work area on cervical-cancer precursor-lesion detection with a cervicogram with two-channel inspection (acetic acid and Lugol’s iodine) [27,28] and one-channel inspection (only acetic acid) [29,30]. As shown in Table 5, the summary of our benchmarking is as follows:
  • We used a high confidence value of 0.9 and IoU values ranging from 0.5 to 0.7 to ensure that our proposed model provided good results. However, our results still outperformed existing studies using SVM [27], Faster-RCNN [28], and a CNN classifier [23,30]. Two cases used fusion inspection with acetic acid and Lugol’s iodine, but our model only used acetic acid. For the entire IoU value, sensitivity performance was 100%, which means that there were no unclassified results or zero false-negative results.
  • HLD-Net developed a detection model with a Faster-RCN architecture with dual-channel fusing detection information across acetic acid and Lugol’s iodine cervicograms [28]; however, this model does not have segmentation capabilities in an infected lesion. AW lesions are hard to detect because their region should intersect with the SCJ in the CA. Using Faster-RCNN achieved sensitivity, specificity, and accuracy of only 40%, 99%, and 69%, respectively. However, our mask-RCNN model was successful in detecting and classifying AW lesions with 98% confidence, 0.5 IoU, 100% sensitivity, 92% specificity, and 96% accuracy to distinguish between normal (without AW lesions) and benign (with AW lesions) tissues.
  • Colposcopic images were developed using the SVM algorithm [27] to classify cervicograms on the basis of the corresponding pathology for visual inspection with acetic acid, visual inspection with Lugol’s iodine, and a combination of the two contrasts. However, SVM only achieved 81.3%, 78.6%, and 80% for sensitivity, specificity, and accuracy, respectively.
  • The combination of the k-means clustering algorithm and CNN classifier was proposed in [29], and the classification result achieved good performance with 86% accuracy. In [23,30], the CNN based on ResNet 50 architecture was used as the classifier. In [23], it achieved 89% sensitivity and 91% accuracy, while in [30], the accuracy, sensitivity, and specificity were 84.10%, 85.38%, and 82.62%, respectively. However, the CNN classifier only predicts normal and abnormal images, and the position of an infected lesion by HPV cannot be detected. This should be improved by the localization and visualization of AW lesions, so that the results can be explained. Our proposed could perform classification, segmentation, and detection tasks, which means that the results are more reliable for clinical practice.
Nonetheless, the current study has some limitations. First, the 484 pathological still images used in model training, validation, and testing were not enough to ensure that this algorithm was suitable for practical clinical use. A well-designed study using the full cervical-cancer precursor-lesion images of many cases should be conducted to verify the clinical accuracy of the proposed model before the mask-RCNN-based combined model is used in clinical practice. Second, abnormal images present three conditions of CIN, including a precancerous condition in which abnormal cells grow on the surface of the cervix. However, because of image limitations, our models could not be trained on different CIN levels. If sufficient image acquisition was achieved for such lesions, the model could be trained and merged into the existing combined model.

4. Conclusions

Deep learning is a data-hungry method, but we showed that a surprisingly small number of cervicogram images could be used to significantly boost diagnosis from what is commonly found in practice. We experimented by selecting the input data according to clinical recommendations for distinguishing anatomy landmarks on the basis of Herf’s theory of the pathogenesis of cervical cancer caused by HPV. This strategy allowed for us to reduce the input data size in our diagnostic model, thereby achieving computational efficiency. Prediction efficiency is key to translating the current study into real-world, resource-poor settings. We developed and validated the mask-RCNN architecture using an instance segmentation approach for cervical-cancer precursor-lesion detection with cervicograms and pathogenesis, which distinguished abnormal cervicograms into VIA+ and VIA−. The proposed model provided effective detection and classified each type of lesion separately with high accuracy. Furthermore, this model demonstrated improved detection ability with a single channel of only acetic acid when compared with another model with a corresponding two-channel pathology for visual inspection with acetic acid and Lugol’s iodine. Our model provides a more reliable lesion detection approach for real-world clinical practice. We look forward to testing and refining these models in larger populations to achieve the more accurate early detection of cervical cancer by medical workers, especially in LMICs.

Author Contributions

P.A., funding acquisition, formal analysis, resources, and writing of the original draft, and review and editing; S.N., conceptualization, data collection, data curation, resources, investigation, supervision, data interpretation, and writing of the original draft; L.N., medical data verification and supervision; I.I. medical data verification and supervision; R.S., medical data verification; L.L., medical data verification; G.O.F., software analyst and review and editing, M.N.R.; software and data analyst; A.I.S., data interpretation; A.D., review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a National Competitive Basic Research Grant from the Ministry of Education, Culture, Research, and Technology, Indonesia, no. 142/E5/PG.02.00.PT/2022 and the Faculty of Medicine, Universitas Sriwijaya, Indonesia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data is available by request.

Acknowledgments

We want to thank the Intelligent System Research Group (ISysRG) Universitas Sriwijaya, Indonesia, for their full support of our research infrastructure.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bray, F.; Ferlay, J.; Soerjomataram, I. Global cancer statistics 2018: Globocan. Estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. International Agency for Research on Cancer (WHO). Indonesia—Global Cancer Observatory 2018. Available online: https://gco.iarc.fr/today/data/factsheets/populations/360-indonesia-fact-sheets.pdf (accessed on 21 February 2022).
  3. Herfs, M.; Soong, T.R.; Delvenne, P.; Crum, C.P. Deciphering the multifactorial susceptibility of mucosal junction cells to HPV Infection and related carcinogenesis. Viruses 2017, 9, 85. [Google Scholar] [CrossRef] [Green Version]
  4. Herfs, M.; Vargas, S.O.; Yamamoto, Y.; Howitt, B.E.; Nucci, M.R.; Hornick, J.L. A novel blueprint for ‘top down’ differentiation defines the cervical squamocolumnar junction during development, reproductive life, and neoplasia. J. Pathol. 2013, 229, 460–468. [Google Scholar] [CrossRef]
  5. Mirkovic, J.; Howitt, B.E.; Roncarati, P.; Philippe, D.; Delvenne, P.; Crum, C.P.; Herfs, M.; Roncarati, P.; Demoulin, S.; Suarez-Carmona, M.; et al. Carcinogenic HPV infection in the cervical squamo-columnar junction. J. Pathol. 2015, 236, 265–271. [Google Scholar] [CrossRef] [PubMed]
  6. Schiffman, M.; Doorbar, J.; Wentzensen, N.; Monk, B.J.; Stanley, M.A.; Franceschi, S. Carcinogenic human papillomavirus infection. Nat. Rev. Dis. Primers 2016, 2, 1–20. [Google Scholar] [CrossRef] [PubMed]
  7. Kundrod, K.A.; Smith, C.A.; Hunt, B.; Schwarz, R.A.; Schmeler, K.; Richards-Kortum, R. Diagnostics advances in technologies for cervical cancer detection in low-resource settings. Expert Rev. Mol. Diagn. 2019, 19, 695–714. [Google Scholar] [CrossRef]
  8. Asiaf, A.; Ahmad, S.T.; Mohammad, S.O.; Zargar, M.A. Review of the current knowledge on the epidemiology, pathogenesis, and prevention of human papillomavirus infection. Eur. J. Cancer Prev. 2007, 23, 206–224. [Google Scholar] [CrossRef] [PubMed]
  9. Gómez, D.T.; Santos, J.L. Human papilloma virus infection and cervical cancer: Pathogenesis and epidemiology. Commun. Curr. Res. Educ. Top. Trends Appl. Microbiol. 2007, 1, 680–688. [Google Scholar]
  10. Silkensen, S.L.; Schiffman, M.; Sahasrabuddhe, V.; Flanigan, S. Is it time to move beyond visual inspection with acetic acid for cervical cancer screening? What is the role of persistent HPV. Glob. Health Sci. Pr. 2018, 6, 242–246. [Google Scholar] [CrossRef] [Green Version]
  11. World Health Organization. WHO Guidelines for Screening and Treatment of Precancerous Lesions for Cervical Cancer Prevention. 2013. Available online: https://apps.who.int/iris/bitstream/handle/10665/96735/WHO_RHR_13.21_eng.pdf?sequence=1. (accessed on 21 February 2022).
  12. WHO. WHO Global World Health Assembly Adopts Global Strategy to Accelerate Cervical Cancer Elimination. 2020. Available online: https://www.who.int/news/item/19-08-2020-world-health-assembly-adopts-global-strategy-to-accelerate-cervical-cancer-elimination (accessed on 29 October 2020).
  13. Mustafa, R.A.; Santesso, N.; Khatib, R.; Mustafa, A.A.; Wiercioch, W.; Kehar, R.; Gandhi, S.; Chen, Y.; Cheung, A.; Hopkins, J.; et al. Systematic reviews and meta-analyses of the accuracy of HPV tests, visual inspection with acetic acid, cytology, and colposcopy. Int. J. Gynecol. Obstet. 2015, 132, 259–265. [Google Scholar] [CrossRef] [Green Version]
  14. Smith, R.A.; Andrews, K.S.; Brooks, D.; Fedewa, S.A.; Manassaram-Baptiste, D.; Saslow, D.; Brawley, O.W.; Wender, R.C. Cancer screening in the United States, 2017: A review of current American Cancer Society guidelines and current issues in cancer screening. CA Cancer J. Clin. 2017, 67, 100–121. [Google Scholar] [CrossRef] [PubMed]
  15. İlkay, Ç. The challenge of pathological diagnosis for precancerous cervical lesions. Middle Black Sea J. Health Sci. 2019, 5, 39–45. [Google Scholar] [CrossRef] [Green Version]
  16. Sauvaget, C.; Fayette, J.; Muwonge, R.; Wesley, R.; Sankaranarayanan, R. Accuracy of visual inspection with acetic acid for cervical cancer screening. Int. J. Gynecol. Obstet. 2011, 113, 14–24. [Google Scholar] [CrossRef] [PubMed]
  17. Shen, D.; Wu, G.; Suk, H. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [Green Version]
  18. Fernandes, K.; Chicco, D.; Cardoso, J.S.; Fernandes, J. Supervised deep learning embeddings for the prediction of cervical cancer diagnosis. PeerJ Comput. Sci. 2018, 4, e154. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Blumenthala, P.D.; Lauterbach, T.M.; Sellors, J.W.; Sankaranarayanan, R. Training for cervical cancer prevention programs in low-resource settings: Focus on visual inspection with acetic acid and cryotherapy. Int. J. Gynecol. Obstet. 2005, 89, S30–S37. [Google Scholar] [CrossRef]
  20. Singh, S.K.; Goyal, A. Performance analysis of machine learning algorithms for cervical cancer detection. Int. J. Healthc. Inf. Syst. Inform. 2020, 15, 1–21. [Google Scholar] [CrossRef]
  21. Cho, B.J.; Choi, Y.J.; Lee, M.J.; Kim, J.H.; Son, G.-H.; Park, S.-H.; Kim, H.-B.; Joo, Y.-J.; Cho, H.-Y.; Kyung, M.S.; et al. Classification of cervical neoplasms on colposcopic photography using deep learning. Sci. Rep. 2020, 10, 1–10. [Google Scholar] [CrossRef]
  22. Good, G. AI Approach Outperformed Human Experts in Identifying Cervical Precancer. 2019. Available online: https://www.nih.gov/news-events/news-releases/ai-approach-outperformed-human-experts-identifying-cervical-precancer (accessed on 21 February 2022).
  23. Park, Y.R.; Kim, Y.J.; Ju, W.; Nam, K.; Kim, S.; Kim, K.G. Comparison of machine and deep learning for the classification of cervical cancer based on cervicography images. Sci. Rep. 2021, 11, 16143. [Google Scholar] [CrossRef]
  24. Rezvy, S.; Zebin, T.; Braden, B.; Pang, W.; Taylor, S.; Gao, X.W. Transfer learning for endoscopy disease detection & segmentation with mask-RCNN benchmark architecture. In Proceedings of the 2nd International Workshop and Challenge on Computer Vision in Endoscopy, EndoCV@ISBI 2020, Iowa City, IA, USA, 3 April 2020; Volume 2595, pp. 68–72. [Google Scholar]
  25. Nurmaini, S.; Rachmatullah, M.N.; Sapitri, A.I.; Darmawahyuni, A.; Jovandy, A.; Firdaus, F.; Tutuko, B.; Passarella, R. Accurate detection of septal defects with fetal ultrasonography images using deep learning-based multiclass instance segmentation. IEEE Access 2020, 8, 196160–196174. [Google Scholar] [CrossRef]
  26. Nurmaini, S.; Rachmatullah, M.N.; Sapitri, A.I.; Darmawahyuni, A.; Tutuko, B.; Firdaus, F.; Partan, R.U.; Bernolian, N. Deep learning-based computer-aided fetal echocardiography: Application to heart standard view segmentation for congenital heart defects detection. Sensors 2021, 21, 8007. [Google Scholar] [CrossRef] [PubMed]
  27. Asiedu, M.N.; Simhal, A.; Chaudhary, U.; Mueller, J.L.; Lam, C.T.; Schmitt, J.W.; Venegas, G.; Sapiro, G.; Ramanujam, N. Development of algorithms for automated detection of cervical pre-cancers with a low-cost, point-of-care, pocket colposcope. IEEE Trans. Biomed. Eng. 2018, 66, 2306–2318. [Google Scholar] [CrossRef] [PubMed]
  28. Yan, L.; Song, H.; Guo, Y.; Ren, P.; Zhou, W.; Li, S.; Yang, J.; Shen, X. HLDnet: Novel deep learning based artificial intelligence tool fuses acetic acid and Lugol’s iodine cervicograms for accurate pre-cancer screening. Biomed. Signal Process. Control 2022, 1, 103163. [Google Scholar] [CrossRef]
  29. Peng, G.; Dong, H.; Liang, T.; Li, L.; Liu, J. Diagnosis of cervical precancerous lesions based on multimodal feature changes. Comput. Biol. Med. 2021, 130, 104209. [Google Scholar] [CrossRef] [PubMed]
  30. Yuan, C.; Yao, Y.; Cheng, B.; Cheng, Y.; Li, Y.; Li, Y.; Liu, X.; Cheng, X.; Xie, X.; Wu, J.; et al. The application of deep learning based diagnostic system to cervical squamous intraepithelial lesions recognition in colposcopy images. Sci. Rep. 2020, 10, 11639. [Google Scholar] [CrossRef]
Figure 1. The proposed methodology of instance segmentation for automate screening of VIA cervicograms based on cervical anatomy with CA and AW lesions.
Figure 1. The proposed methodology of instance segmentation for automate screening of VIA cervicograms based on cervical anatomy with CA and AW lesions.
Sensors 22 05489 g001
Figure 2. Cervicogram with adequate anatomy landmark. (a) Normal condition with the CA, SCJ, and TZ; (b) SC junction cells to immature metaplasia; (c) abnormal condition with AW lesion intersected with the SCJ.
Figure 2. Cervicogram with adequate anatomy landmark. (a) Normal condition with the CA, SCJ, and TZ; (b) SC junction cells to immature metaplasia; (c) abnormal condition with AW lesion intersected with the SCJ.
Sensors 22 05489 g002
Figure 3. Sample of annotated cervicograms by gynecological oncologist clinicians for standard cervicogram view in (a) normal with a red line as the SCJ and red area as the CA; (b) abnormal with a green line as the AW lesion and red line as the SCJ.
Figure 3. Sample of annotated cervicograms by gynecological oncologist clinicians for standard cervicogram view in (a) normal with a red line as the SCJ and red area as the CA; (b) abnormal with a green line as the AW lesion and red line as the SCJ.
Sensors 22 05489 g003
Figure 4. Sample of annotated cervicograms by gynecological oncologists for AW lesion detection (precursor cancer lesion) and normal cervicograms. In the annotation, the region with the red line is the CA, and with the green line is the AW lesion; (a) raw data; (b) annotation label, and (c) squamocolumnar junction (SCJ) forms.
Figure 4. Sample of annotated cervicograms by gynecological oncologists for AW lesion detection (precursor cancer lesion) and normal cervicograms. In the annotation, the region with the red line is the CA, and with the green line is the AW lesion; (a) raw data; (b) annotation label, and (c) squamocolumnar junction (SCJ) forms.
Sensors 22 05489 g004
Figure 5. RPNs with the ResNet 50 backbone and FCN architecture for AW lesion detection.
Figure 5. RPNs with the ResNet 50 backbone and FCN architecture for AW lesion detection.
Sensors 22 05489 g005
Figure 6. An example of a feature map extracted from the ResNet50 backbone in the RPN.
Figure 6. An example of a feature map extracted from the ResNet50 backbone in the RPN.
Sensors 22 05489 g006
Figure 7. Training and validation loss from the learning process with the ResNet50 backbone.
Figure 7. Training and validation loss from the learning process with the ResNet50 backbone.
Sensors 22 05489 g007
Figure 8. Classification performance with confusion matrix and ROC curve for three backbones.
Figure 8. Classification performance with confusion matrix and ROC curve for three backbones.
Sensors 22 05489 g008
Figure 9. The result of mask-RCNN to identify CA and AW lesions. (left to right) Raw image, annotation image (red line for CA and green line for AW lesions in the ground truth), and prediction of CA and AW lesions (red for CA, and green for AW lesions). (ac) Abnormal cervicograms; (df) normal cervicograms.
Figure 9. The result of mask-RCNN to identify CA and AW lesions. (left to right) Raw image, annotation image (red line for CA and green line for AW lesions in the ground truth), and prediction of CA and AW lesions (red for CA, and green for AW lesions). (ac) Abnormal cervicograms; (df) normal cervicograms.
Sensors 22 05489 g009aSensors 22 05489 g009b
Table 1. Data distribution for the learning process.
Table 1. Data distribution for the learning process.
CervicogramTrainingValidationTestingTotal
Normal1871524226
Abnormal2062631263
Table 2. Mask-RCNN performance with different learning rates.
Table 2. Mask-RCNN performance with different learning rates.
Learning RateIoU (%)DSC (%)mAP (%)
CAAW LesionCAAW LesionCAAW Lesion
0.00165.6557.2873.1976.9484.1799.83
0.000163.6172.4372.5588.8186.90100
0.0000137.9827.8151.8553.3775.4295.60
Table 3. Mask-RCNN performance with three backbone architectures.
Table 3. Mask-RCNN performance with three backbone architectures.
ArchitectureIoU (%)DSC (%)mAP (%)
CA RegionAW LesionCA RegionAW LesionCA RegionAW Lesion
ResNet5063.6172.4372.5588.8186.90100
ResNet10163.7372.7373.2286.7383.7599.85
MobileNetV162.3871.0966.8785.0670.5972.09
Table 4. Mask-RCNN classification performance with three backbone architectures.
Table 4. Mask-RCNN classification performance with three backbone architectures.
ArchitecturePerformance (%)
AccuracySensitivitySpecificityPrecisionF1 Score
ResNet5096.291009293.5496.67
ResNet10189.101008080.6489.29
MobileNetV156.361005022.5836.84
Table 5. Benchmarking results with the existing research on cervical-cancer precursor lesions.
Table 5. Benchmarking results with the existing research on cervical-cancer precursor lesions.
MethodsLearning ProcessIoUSensitivitySpecificityAccuracyConfidenceInspection
SVM [27]Classification-0.810.790.80-Acetic acid and Lugol’s iodine
Faster-RCNN [28]Classification and detection0.20.820.900.860.80Acetic acid and Lugol’s iodine
0.30.630.940.780.80
0.40.400.990.690.80
0.40.550.670.610.80Acetic acid
0.40.490.570.53080Lugol’s iodine
K-means clustering and CNNs classifier [29]Classification-0.840.900.86-Acetic acid
CNN with ResNet 50 [23]Classification-0.89-0.91-Acetic acid
CNN with ResNet 50 [30]Classification-0.8582.620.84-Acetic acid and Lugol’s iodine
Mask-RCNN with ResNet50 (our model)Segmentation, classification, and detection0.410.920.960.97Acetic acid
0.510.920.960.97
0.610.770.870.98
0.710.500.600.98
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Agustiansyah, P.; Nurmaini, S.; Nuranna, L.; Irfannuddin, I.; Sanif, R.; Legiran, L.; Rachmatullah, M.N.; Florina, G.O.; Sapitri, A.I.; Darmawahyuni, A. Automated Precancerous Lesion Screening Using an Instance Segmentation Technique for Improving Accuracy. Sensors 2022, 22, 5489. https://doi.org/10.3390/s22155489

AMA Style

Agustiansyah P, Nurmaini S, Nuranna L, Irfannuddin I, Sanif R, Legiran L, Rachmatullah MN, Florina GO, Sapitri AI, Darmawahyuni A. Automated Precancerous Lesion Screening Using an Instance Segmentation Technique for Improving Accuracy. Sensors. 2022; 22(15):5489. https://doi.org/10.3390/s22155489

Chicago/Turabian Style

Agustiansyah, Patiyus, Siti Nurmaini, Laila Nuranna, Irfannuddin Irfannuddin, Rizal Sanif, Legiran Legiran, Muhammad Naufal Rachmatullah, Gavira Olipa Florina, Ade Iriani Sapitri, and Annisa Darmawahyuni. 2022. "Automated Precancerous Lesion Screening Using an Instance Segmentation Technique for Improving Accuracy" Sensors 22, no. 15: 5489. https://doi.org/10.3390/s22155489

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop