Next Article in Journal
Visual Scoring of Sacroiliac Joint/Sacrum Ratios of Single-Photon Emission Computed Tomography/Computed Tomography Images Affords High Sensitivity and Negative Predictive Value in Axial Spondyloarthritis
Next Article in Special Issue
Brain Tumor Class Detection in Flair/T2 Modality MRI Slices Using Elephant-Herd Algorithm Optimized Features
Previous Article in Journal
Deep Learning-Based Recognition of Cervical Squamous Interepithelial Lesions
Previous Article in Special Issue
DeepTumor: Framework for Brain MR Image Classification, Segmentation and Tumor Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Applying Deep Transfer Learning to Assess the Impact of Imaging Modalities on Colon Cancer Detection

Department of Computer Science, King Abdulaziz University, Jeddah 21589, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(10), 1721; https://doi.org/10.3390/diagnostics13101721
Submission received: 1 April 2023 / Revised: 4 May 2023 / Accepted: 11 May 2023 / Published: 12 May 2023
(This article belongs to the Special Issue AI as a Tool to Improve Hybrid Imaging in Cancer—2nd Edition)

Abstract

:
The use of medical images for colon cancer detection is considered an important problem. As the performance of data-driven methods relies heavily on the images generated by a medical method, there is a need to inform research organizations about the effective imaging modalities, when coupled with deep learning (DL), for detecting colon cancer. Unlike previous studies, this study aims to comprehensively report the performance behavior for detecting colon cancer using various imaging modalities coupled with different DL models in the transfer learning (TL) setting to report the best overall imaging modality and DL model for detecting colon cancer. Therefore, we utilized three imaging modalities, namely computed tomography, colonoscopy, and histology, using five DL architectures, including VGG16, VGG19, ResNet152V2, MobileNetV2, and DenseNet201. Next, we assessed the DL models on the NVIDIA GeForce RTX 3080 Laptop GPU (16GB GDDR6 VRAM) using 5400 processed images divided equally between normal colons and colons with cancer for each of the imaging modalities used. Comparing the imaging modalities when applied to the five DL models presented in this study and twenty-six ensemble DL models, the experimental results show that the colonoscopy imaging modality, when coupled with the DenseNet201 model under the TL setting, outperforms all the other models by generating the highest average performance result of 99.1% (99.1%, 99.8%, and 99.1%) based on the accuracy results (AUC, precision, and F1, respectively).

1. Introduction

Cancer is an oftentimes rapidly spreading disease that drastically affects human health [1]. One of the most common types of cancer is colon cancer, which is sometimes caused by polyps in the colon wall, as shown in Figure 1 [2]. Colon cancer is the second and third most prevalent cancer in terms of death and incidence rates, respectively [3]. Consequently, previous studies have proposed many methods for improving the detection of colon cancer [4,5,6,7]. Medical imaging is one method used for detecting colon cancer. However, dealing with large numbers of medical images causes difficulties for specialists, which results in delays in the detection of colon cancer and, thus, delays in treatment. Therefore, automating the detection of colon cancer using deep learning (DL) attends to these challenges effectively.
Patino-Barrientos et al. [8] employed the VGG16 DL model to classify colon polyps as either malignant or nonmalignant, using an image dataset that consisted of 600 colonoscopy-derived images (from a private institution) based on Kudo’s method. The VGG16 model was utilized in two different ways, as demonstrated as follows: for the first situation, a pre-trained VGG16 model on ImageNet was used for the feature extraction, freezing layers pertaining to the convolutional feature extraction while changing the densely connected classifier to address the new binary classification problem. The convolutional feature extraction of the VGG16 model was then applied to the training data related to colon cancer, and the resulting features were used as inputs for the densely connected classifier to induce the model to perform predictions of unseen colon cancer images. For the second situation, a pre-trained VGG16 with fine tuning was used, which freezes the bottom layers while unfreezing the remaining layers. Compared to other machine-learning-based methods that use hand-crafted features of histograms of an oriented gradient, the results from a testing subset of the dataset demonstrate the superiority of the VGG16 model. Sarwinda et al. [9] aimed to classify colon cancer as malignant or benign using histology-based images. They utilized ResNet-based DL models, namely ResNet-18 and ResNet-50. Their approach worked as follows: firstly, the images were pre-processed using the contrast limited adaptive histogram equalization technique to generate improved images. Then, employing ResNet-18 and ResNet-50 and using feature extraction, they froze all the layers except for the densely connected classifier to deal with the binary classification problem. Features were extracted from pre-processed training images and given to the densely connected classifier, which was followed by the performing of predictions on the testing images. In terms of the evaluation, the dataset was divided into training and testing three times according to user-specified percentages. The reported results demonstrated the feasibility of ResNet-based DL models.
Ponzio et al. [10] aimed to classify colon cancer based on histological images. They utilized the VGG16 DL model in three different ways, including transfer learning. The first suggested model (a fully trained VGG16) consisted of training a VGG16 from scratch on colon cancer data and performing predictions of unseen histology-based images that were related to colon cancer. The second suggested model (a pre-trained VGG16 model on ImageNet with feature extraction) was applied to histology-based training images to extract features, and provided with corresponding labels for the machine learning algorithm (SVM). The induced SVM model was then applied to a testing set (consisting of feature vectors constructed from a pre-trained VGG16) to generate predictions. The third suggested model (a pre-trained VGG16 model on ImageNet with fine tuning) froze some layers while unfreezing the remaining layers. The experimental results demonstrated that pre-trained VGG16 models utilizing transfer learning (i.e., the second and third VGG16 models) outperformed the supervised learning approach of the VGG16 which was fully trained from scratch. Basha et al. [11] developed a CNN called RCCNet to classify colon cancer nuclei into four categories: miscellaneous, fibroblast, epithelial, and inflammatory. Their developed model was compared with various DL models: WRN, GoogLeNet, AlexNet, softmaxCNN, and softmaxCNN_IN27, and their proposed model achieved the best performance results. Ribeiro et al. [12] used CNN with data augmentation to classify colon images into two classes: healthy and abnormal. The experimental results demonstrated the good performance of the utilized CNN.
The problem with detecting colon cancer using medical images depends on the data-driven methods used and the images generated by an imaging modality. Unlike previous studies that have focused on evaluating the performance behavior of DL models in terms of detecting colon cancer [8,9,10,11,12], our contributions can be summarized as follows:
(1)
We utilized three imaging modalities, namely, CT [13], colonoscopy [14], and histology [15,16], with five DL architectures, including VGG16 [17], VGG19 [17], ResNet152V2 [18], MobileNetV2 [19], and DenseNet201 [20].
(2)
We comprehensively reported the performance behavior for the detection of colon cancer, including generated images via different modalities coupled with DL models in the transfer learning setting. Moreover, we constructed 26 ensemble DL models and compared their performance against the 5 studied DL models.
(3)
We identified the best overall imaging modality and DL model for the detection of colon cancer. Specifically, our results reported that colonoscopy-based images outperformed CT-based (and histology-based) images when coupled with DL models.
(4)
Our reported results demonstrate the superiority of DenseNet201 compared to 30 other DL models, including 4 DL methods and 26 ensemble DL models. According to the average performance results, measured using a 5-fold cross-validation of the whole dataset of colonoscopy-based colon cancer images, DenseNet201 generated the highest average accuracy of 99.1%, the highest average area under the ROC curve of 99.1%, the highest average F1 of 99.1%, and the highest average precision of 99.8%. Since the 26 ensemble DL models generated inferior performance results, we moved their results into the Supplementary Materials File.

2. Materials and Methods

2.1. Datasets

This study used four publicly available datasets for detecting colon cancer. Firstly, we used the Cancer Genome Atlas Colon Adenocarcinoma Collection (TCGA-COAD) dataset of CT imaging modalities (accessible at https://doi.org/10.7937/K9/TCIA.2016.HJJHBOXZ accessed on 6 January 2023), which includes 8387 CT images of colon cancer [21,22,23]. Secondly, we used the CT COLONOGRAPHY dataset of CT imaging modalities (accessible at https://doi.org/10.7937/K9/TCIA.2015.NWTESAY1 accessed on 6 January 2023), which includes 941,771 CT images, 268,652 of which are relevant to the current field of study [24,25,26]. Thirdly, we used the HyperKvasir Dataset of colonoscopy imaging modalities (accessible at https://doi.org/10.17605/OSF.IO/MH9SJ accessed on 6 January 2023), which includes 10,662 images and 374 videos that represent 23 and 30 categories, respectively, and 99,417 undefined images. Among the identified dataset, there are four videos of an instance of colon cancer and one video of a normal colon [27,28]. Fourthly, we used the NCT-CRC-HE-100K-NONORM Dataset of histology imaging modalities (accessible at https://search.datacite.org/works/10.5281/zenodo.1214456 accessed on 6 January 2023), which includes 100,000 histology images and 23,080 images related to our study, which were divided into 14,317 images of instances of colon cancer and 8763 images of normal colons [29].

2.2. Pre-Processing

Pre-processing is a necessary phase of a medical image. It significantly affects the prediction results for colon cancer [30]. The datasets were obtained from various sources and techniques, including a subset of videos and poor-quality images with highlighted information, black borders, blurred contrast, and noise, which could influence the learning and prediction of the model. Therefore, we applied pre-processing to clean datasets, enhanced medical image conversion, generated a dataset of images from the videos, deleted blurred colon images, improved image quality, removed unwanted objects, and balanced class distribution. Firstly, we cleaned the datasets of lesions that were unrelated to our study. Then, we generated an image dataset for the colonoscopy technique by extracting frames from videos of the HyperKvasir dataset depending on FPS [31]. Thereafter, we removed the highlighted information by converting the color colon images to grayscale and using the THRESH_BINARY method to generate a binary mask and distinguish high and low pixel values; this was followed by the inpainting technique, which reconstructs the colon image using nearby pixels [32,33]. Next, we processed highly unbalanced datasets using a random undersampling technique that randomly selects samples from the majority class to equate to the minority classes [34]. Table 1 shows the number of images used for detecting colon cancer after applying the random undersampling method. Additionally, we enhanced the contrast of images using the CLAHE method followed by the Gaussian blur technique to remove any noise that the CLAHE method may have caused [35,36,37]. Moreover, we removed the black borders from the images to focus on processing the important features [38]. Finally, we changed the multiscale of images to fit the inputs of the CNN models using the INTER_LINEAR technique to 224 × 224 [39]. Figure 2 shows images of the colon before and after the pre-processing procedure.

2.3. Deep Learning Approach

The DL approach used for predicting colon cancer and distinguishing between normal colons (negative) and colon cancer (positive) is shown in Figure 3. S = x i , y i m i = 1 is a training set that includes m-labeled images obtained from various imaging modalities. Each training example has a class label (0 or 1), where 0 indicates a normal colon and 1 indicates colon cancer. This study used five pre-trained CNN models, including VGG16, VGG19, ResNet152V2, MobileNetV2, and DenseNet201. We adapted the five DL models to our problem using a transfer learning method based on the ImageNet dataset and feature extraction technique [40], whereby all layers were frozen with weights of ImageNet except for the last layer, which was replaced by a new dense layer that had one neuron and sigmoid activation and was trained independently on each of the colon cancer datasets, as shown in Figure 4. Each of the five DL models were trained independently on processed images of a given modality. Then, the unseen datasets were tested on the trained models of the same modality to generate predictions mapped to 0 and 1 as follows: if the prediction is greater than 0.5, it is set to 1, which thus indicates colon cancer; otherwise, it indicates a normal colon.

3. Results

3.1. Classification Methodology

For each image dataset, we investigated the performance of three imaging modalities (CT, histology, and colonoscopy) through five DL models (VGG16, VGG19, ResNet152V2, MobileNetV2, and DenseNet201) for predicting colon cancer. The five DL models were utilized in the transfer learning setting to address the classification task. After training, the DL models were applied to the testing images to generate predictions, which were mapped according to the following specified thresholds: 0 (normal colon) or 1 (colon cancer). Furthermore, we constructed 26 ensemble DL models. Since the 26 ensemble DL models did not outperform DenseNet201, we recorded their results in the Supplementary Materials File. To evaluate the performance of the models, we used five performance metrics: accuracy (ACC), precision (PRE), recall (REC), F1, and area under the ROC curve (AUC) [40,41]. To validate the performance of the DL models over the entire dataset, we applied a five-fold cross-validation by partitioning each dataset into five folds. For each run, we assigned five folds: four for the training set and one for the test set, where the prediction was applied to the testing fold. Finally, we reported the average performance results of the five runs using the following performance metrics:
            A C C = T P + T N T P + T N + F P + F N
P R E = T P T P + F P
R E C = T P T P + F N
              F 1 = 2 P R E R E C P R E + R E C
where TP stands for true positive, referring to the number of colon cancer images that were correctly classified as colon cancer. FN stands for false negative, referring to the number of colon cancer images that were incorrectly classified as a normal colon. TN stands for true negative, referring to the number of normal colon images that were correctly classified as a normal colon. FP stands for false positive, referring to the number of normal colon images that were incorrectly classified as colon cancer.

3.2. Implementation Details

In this experiment, we used the Spyder editor (Version 4.2.5), which we accessed using Anaconda (Version 4.12.0) in Python (Version 3.8.8) [42,43]. We used the Keras library to run five DL models [44]. The datasets were processed in the pre-processing stage using OpenCV and NumPy libraries [45,46]. The training and testing of the DL models were conducted on the NVIDIA GeForce RTX 3080 Laptop GPU with 16 GB GDDR6 VRAM. For assessing the five DL models, we used the Sklearn library [46]. To obtain the box plot statistics for the training and testing phases, we utilized ggplot2 in R [47].

3.3. Classification Results

The datasets used in this study included 5400 processed images that were divided equally between normal colon and colon cancer and related to three types of medical images. Based on that, we assessed the image datasets obtained from three imaging modalities using five DL models (and we moved twenty-six ensemble DL models to the Supplementary Materials File because they produced inferior results), which was then followed by reporting their performances using a five-fold cross-validation.

3.3.1. Training Results

Figure 5 illustrates the performance of the DL models when applied to images derived from imaging modalities on the training sets during a five-fold cross-validation based on the ACC, PRE, REC, and F1 performance measurements. The boxplots showed that DenseNet201 generated the highest performance results, according to ACC and PRE, when coupled with images derived from colonoscopy and CT imaging modalities. When DenseNet201 was coupled with images derived from CT imaging modality, it generated the highest results. The DL models achieved poor performance results when they were coupled with images that were derived from a histology imaging modality.

3.3.2. Testing Results

Figure 6 shows that DenseNet201 achieved the best average performance results when coupled with images that were derived from colonoscopy and CT imaging modalities. Specifically, DenseNet201 (when coupled with colonoscopy-based images) achieved 99.1% (99.8% and 99.1%) according to ACC (PRE and F1, respectively). Moreover, it obtained the best average REC of 99.4% for images that were derived from a CT imaging modality, as shown in Table 2. For images derived from a histology imaging modality, MobileNetV2 achieved the lowest average performance results (66.6–71.4%) based on employed performance measures. According to Table 2, the colonoscopy imaging modality, when coupled with the DenseNet201 model, achieved the most reliable performance results. Figure 7 illustrates the combined confusion matrices of a five-fold cross-validation on the test sets. For each DL model and imaging modality, the sum of five test splits corresponds to the combined confusion matrices, and the sum of entries indicates that the whole dataset was used. Figure 8 displays the ROC curves for five DL models applied to the image datasets obtained from CT, histology, and colonoscopy imaging modalities. The DL model with the highest curve indicates the highest AUC results. It can be seen that DenseNet201 archives the highest AUC values, which are recorded in Table 2.

4. Discussion

Our DL system included four parts: (1) data acquisition; (2) data pre-processing; (3) the handling of the issue of binary classification under different medical imaging techniques, where we aimed to detect colon cancer by distinguishing between normal colon and colon cancer; and (4) the investigating of various imaging modalities through different DL models in the transfer learning setting. After the image dataset acquisition, which included 5400 images from normal colon and colon cancer of different imaging modalities, we provided the processed image datasets to DL models and reported the performance results using a five-fold cross-validation.
The technical contributions of this study are as follows: (1) the application of DL models to detect colon cancer under different imaging modalities; (2) the conducting of experimental studies in the transfer learning setting using processed datasets of 5400 images (900 of normal colons and 900 of colon cancer for computed tomography images; 900 of normal colons and 900 of colon cancer for histology images; and 900 of normal colons and 900 of colon cancer for standard colonoscopy images); (3) the inclusion of an extensive performance comparison of 5 DL models and 26 ensemble methods; and (4) the identification of the best DL model associated with images generated by an imaging modality.
For an explanation pertaining to transfer learning, we passed the colon cancer image samples through the feature extraction part of a pre-trained CNN on ImageNet to extract the features, which were provided to a new densely connected classifier that was trained from scratch. In other words, we reused the feature extraction part of a pre-trained CNN on ImageNet by freezing the involved layers to extract the features from colon cancer images while changing the densely connected classifier of the pertained CNN on ImageNet to address the binary class classification problem in this study. It is worth noting that the term ‘feature extraction part’ refers to layers in the CNN that are related to feature extraction, such as convolutional and pooling layers. Additionally, freezing a layer prevents its weight from being updated [48]. It is evident that transfer learning is attributed to the weights kept in the feature extraction part of the pre-trained CNN.
In this study, we employed deep transfer learning models to (1) report the performance behavior of DL models when coupled with images generated via studied imaging modalities; (2) assess the feasibility of DL; and (3) promote the use of AI as a tool that can help doctors in the detection of colon cancer by identifying which imaging modality leads to high performance results when coupled with a DL model. All the studied datasets, which are cited in the datasets subsection, are labeled by domain experts and are publicly available. The colon cancer CT image dataset (and the other colon cancer datasets obtained from different modalities) consisted of 1800 images with a uniform class distribution. For the training phase during a 5-fold cross-validation, we utilized a batch size set to 20 as in [49,50], set the learning rate for the SGD optimizer to 0.0001 as in [51], and used binary_crossentropy as the loss function. Moreover, we trained the models for 20 epochs coinciding with Ref. [50]. We used the testing fold to assess the performance of each trained model. As the five-fold cross-validation ran five times, we reported the average performance on the testing folds. In other words, we utilized the five-fold cross-validation to report the performance on the whole dataset, as combining the images on the five testing folds corresponded to the 1800 images in the colon cancer CT image dataset. It is worth mentioning that during an iteration of a 5-fold cross-validation, the testing fold included 360 images from the 2 categories (180 images from each category), and the training splits included 1440 images from the 2 categories (720 images from each category).
For the ensemble methods, the 26 (i.e., 5 2 + 5 3   + 5 4 +   5 5 ) ensemble DL models were constructed using a majority vote as follows: for an ensemble of two DL models, where 5 2 was the number of different ensemble DL models, each ensemble consisted of two DL models out of five, and n r denoted the binomial theorem. The five DL models included VGG16, VGG19, ResNet152V2, MobileNetV2, and DenseNet201. Therefore, we created ten ensemble DL models by taking combinations of two out of the five DL models. For an ensemble composed of three DL models, 5 3 was the number of different ensemble methods, where each ensemble consisted of three DL models out of five. Therefore, we created ten ensemble DL models by taking combinations of three out of the five DL models. For the ensemble DL models composed of four DL models, 5 4 was the number of different ensemble DL models. Therefore, we created five ensemble DL models by taking combinations of four out of the five DL models. The last ensemble DL model consisted of five DL models. Therefore, we created one (i.e., 5 5 ) ensemble DL model. We used majority vote when making a prediction in each of the 26 ensemble methods. Since all the 26 ensemble DL models did not perform well compared to DenseNet201, we moved their results to the Supplementary Materials File.

5. Conclusions and Future Work

To assess image modalities for the task of colon cancer detection, we proposed using DL models under transfer learning. For the image dataset preparation, we performed the following tasks: cleaning, extracting frames, removing unwanted objects, handling imbalanced categories, image enhancement, noise removal, removing black borders, cropping, and resizing images. Then, several DL models (VGG16, VGG19, ResNet152V2, MobileNetV2, and DenseNet201) were coupled with colon cancer images from various imaging modalities (CT, histology, and colonoscopy) to discriminate between instances of normal colons and colon cancer. Each DL model was independently trained on the colon cancer image datasets of a given modality and then applied to the test set to perform predictions. For an assessment of the DL models, including the 26 ensemble-based DL models, we used a 5-fold cross-validation and several performance measures, including accuracy, precision, recall, and F1. Unlike histology-based (and CT-based) images, the experimental results demonstrated that DenseNet201 (under transfer learning with feature extraction) coupled with images derived from standard colonoscopy achieved the best average accuracy of 99.1%, the best average AUC of 99.1%, the best average precision of 99.8%, and the best average F1 of 99.1%.
Future work in this field should include the following: (1) the utilization of the presented deep transfer learning method to investigate other imaging modalities, such as MRI and PET, coupled with different pre-trained models, and (2) the expansion of the binary classification problem to attend to the multiclass classification problem in order to address classification tasks that are related to different cancer types.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics13101721/s1.

Author Contributions

T.T. and W.A. conceived and designed the study. W.A. performed the analysis. W.A. wrote the manuscript. T.T. and W.A. revised and edited the manuscript. T.T. supervised the study. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The publicly available datasets in this manuscript are as follows: The TCGA-COAD dataset is available at https://doi.org/10.7937/K9/TCIA.2016.HJJHBOXZ, accessed on 6 January 2023. The CT COLONOGRAPHY dataset is available at https://doi.org/10.7937/K9/TCIA.2015.NWTESAY1, accessed on 6 January 2023. The HyperKvasir dataset is available at https://doi.org/10.17605/OSF.IO/MH9SJ, accessed on 6 January 2023. The NCT-CRC-HE-100K-NONORM dataset is available at https://search.datacite.org/works/10.5281/zenodo.1214456, accessed on 6 January 2023.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DL   Deep Learning
CT   Computed Tomography
VGG   Visual Geometry Group
ResNet   Residual Neural Network
CNN   Convolutional Neural Network
SVM   Support Vector Machine
CLAHE   Contrast Limited Adaptive Histogram Equalization
Conv   Convolutional
FC   Fully Connected
CC   Colon Cancer
SGD   Stochastic Gradient Descent
MRI   Magnetic Resonance Imaging
PET   Positron Emission Tomography
ACC   Accuracy
PRE   Precision
REC   Recall
ROC   Receiver Operating Characteristic
AUC   Area Under the ROC Curve

References

  1. Hassin, O.; Nataraj, N.B.; Shreberk-Shaked, M.; Aylon, Y.; Yaeger, R.; Fontemaggi, G.; Mukherjee, S.; Maddalena, M.; Avioz, A.; Iancu, O. Different hotspot p53 mutants exert distinct phenotypes and predict outcome of colorectal cancer patients. Nat. Commun. 2022, 13, 2800. [Google Scholar] [CrossRef] [PubMed]
  2. Mori, G.; Rampelli, S.; Orena, B.S.; Rengucci, C.; De Maio, G.; Barbieri, G.; Passardi, A.; Casadei Gardini, A.; Frassineti, G.L.; Gaiarsa, S. Shifts of faecal microbiota during sporadic colorectal carcinogenesis. Sci. Rep. 2018, 8, 10329. [Google Scholar] [CrossRef] [PubMed]
  3. Chhikara, B.S.; Parang, K. Global Cancer Statistics 2022: The trends projection analysis. Chem. Biol. Lett. 2023, 10, 451. [Google Scholar]
  4. Li, J.N.; Yuan, S.Y. Fecal occult blood test in colorectal cancer screening. J. Dig. Dis. 2019, 20, 62–64. [Google Scholar] [CrossRef] [PubMed]
  5. Carethers, J.M. Fecal DNA testing for colorectal cancer screening. Annu. Rev. Med. 2020, 71, 59–69. [Google Scholar] [CrossRef]
  6. Gluecker, T.M.; Johnson, C.D.; Harmsen, W.S.; Offord, K.P.; Harris, A.M.; Wilson, L.A.; Ahlquist, D.A. Colorectal cancer screening with CT colonography, colonoscopy, and double-contrast barium enema examination: Prospective assessment of patient perceptions and preferences. Radiology 2003, 227, 378–384. [Google Scholar] [CrossRef]
  7. Van Cutsem, E.; Verheul, H.M.; Flamen, P.; Rougier, P.; Beets-Tan, R.; Glynne-Jones, R.; Seufferlein, T. Imaging in colorectal cancer: Progress and challenges for the clinicians. Cancers 2016, 8, 81. [Google Scholar] [CrossRef]
  8. Patino-Barrientos, S.; Sierra-Sosa, D.; Garcia-Zapirain, B.; Castillo-Olea, C.; Elmaghraby, A. Kudo’s classification for colon polyps assessment using a deep learning approach. Appl. Sci. 2020, 10, 501. [Google Scholar] [CrossRef]
  9. Sarwinda, D.; Paradisa, R.H.; Bustamam, A.; Anggia, P. Deep learning in image classification using residual network (ResNet) variants for detection of colorectal cancer. Procedia Comput. Sci. 2021, 179, 423–431. [Google Scholar] [CrossRef]
  10. Ponzio, F.; Macii, E.; Ficarra, E.; Di Cataldo, S. Colorectal cancer classification using deep convolutional networks. In Proceedings of the 11th International Joint Conference on Biomedical Engineering Systems and Technologies, Madeira, Portugal, 19–21 January 2018; pp. 58–66. [Google Scholar] [CrossRef]
  11. Basha, S.S.; Ghosh, S.; Babu, K.K.; Dubey, S.R.; Pulabaigari, V.; Mukherjee, S. Rccnet: An efficient convolutional neural network for histological routine colon cancer nuclei classification. In Proceedings of the 15th International Conference on Control, Automation, Robotics and Vision (ICARCV), Singapore, 18–21 November 2018; pp. 1222–1227. [Google Scholar] [CrossRef]
  12. Ribeiro, E.; Uhl, A.; Häfner, M. Colonic polyp classification with convolutional neural networks. In Proceedings of the 2016 IEEE 29th International Symposium on Computer-Based Medical Systems (CBMS), Belfast and Dublin, Ireland, 20–24 June 2016; pp. 253–258. [Google Scholar] [CrossRef]
  13. Boellaard, T.N. Refining CT colonography methods. Eur. J. Radiol. 2013, 82, 1144–1158. [Google Scholar] [CrossRef]
  14. Schwab, M. Encyclopedia of Cancer; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  15. Sena, P.; Fioresi, R.; Faglioni, F.; Losi, L.; Faglioni, G.; Roncucci, L. Deep learning techniques for detecting preneoplastic and neoplastic lesions in human colorectal histological images. Oncol. Lett. 2019, 18, 6101–6107. [Google Scholar] [CrossRef] [PubMed]
  16. Noorbakhsh, J.; Farahmand, S.; Foroughi Pour, A.; Namburi, S.; Caruana, D.; Rimm, D.; Soltanieh-Ha, M.; Zarringhalam, K.; Chuang, J.H. Deep learning-based cross-classifications reveal conserved spatial behaviors within tumor histological images. Nat. Commun. 2020, 11, 6367. [Google Scholar] [CrossRef] [PubMed]
  17. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
  18. Leibe, B.; Matas, J.; Sebe, N.; Welling, M. Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Proceedings, Part IV; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar] [CrossRef]
  19. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef]
  20. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Proceedings of the IEEE conference on computer vision and pattern recognition. Densely Connect. Convolutional Netw. 2017, 4700–4708. [Google Scholar] [CrossRef]
  21. The Cancer Genome Atlas Program (TCGA). Available online: http://cancergenome.nih.gov/ (accessed on 6 January 2023).
  22. Kirk, S.; Lee, Y.; Sadow, C.A.; Levine, S.; Roche, C.; Bonaccio, E.; Filiippini, J. Radiology data from the cancer genome atlas colon adenocarcinoma [TCGA-COAD] collection. Cancer Imaging Arch. 2016. [Google Scholar] [CrossRef]
  23. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M. The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef]
  24. Smith, K.; Clark, K.; Bennett, W.; Nolan, T.; Kirby, J.; Wolfsberger, M.; Moulton, J.; Vendt, B.; Freymann, J. Data from CT_COLONOGRAPHY. Cancer Imaging Arch. 2015. [Google Scholar] [CrossRef]
  25. Johnson, C.D.; Chen, M.; Toledano, A.Y.; Heiken, J.P.; Dachman, A.; Kuo, M.D.; Menias, C.O.; Siewert, B.; Cheema, J.I.; Obregon, R.G. Accuracy of CT colonography for detection of large adenomas and cancers. N. Engl. J. Med. 2008, 359, 1207–1217. [Google Scholar] [CrossRef]
  26. Borgli, H.; Thambawita, V.; Smedsrud, P.H.; Hicks, S.; Jha, D.; Eskeland, S.L.; Randel, K.R.; Pogorelov, K.; Lux, M.; Nguyen, D.T.D. HyperKvasir, a comprehensive multi-class image and video dataset for gastrointestinal endoscopy. Sci. Data 2020, 7, 283. [Google Scholar] [CrossRef]
  27. The HyperKvasir Dataset. Available online: https://osf.io/mh9sj/ (accessed on 6 January 2023).
  28. Kather, J.N.; Halama, N.; Marx, A. 100,000 Histological Images of Human Colorectal Cancer and Healthy Tissue. Zenodo 2018. [Google Scholar] [CrossRef]
  29. Srivaramangai, R.; Hiremath, P.; Patil, A.S. Preprocessing MRI images of colorectal cancer. Int. J. Comput. Sci. Issues (IJCSI) 2017, 14, 48. [Google Scholar]
  30. Sarraf, S.; Noori, M. Multimodal deep learning approach for event detection in sports using Amazon SageMaker. AWS Mach. Learn. Blog 2021. [Google Scholar]
  31. Minichino, J.; Howse, J. Learning OpenCV 3 Computer Vision with Python; Packt Publishing Ltd.: Birmingham, UK, 2015. [Google Scholar]
  32. Lokhande, D.; Zope, R.G.; Bendre, V.; Kopargaon, S.C. Image Inpainting Image Inpainting. Available online: http://ijcsn.org/IJCSN-2014/3-1/Image-Inpainting.pdf (accessed on 6 January 2023).
  33. Mohammed, R.; Rawashdeh, J.; Abdullah, M. Machine learning with oversampling and undersampling techniques: Overview study and experimental results. In Proceedings of the 2020 11th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 7–9 April 2020; pp. 243–248. [Google Scholar] [CrossRef]
  34. Zuiderveld, K. Contrast limited adaptive histogram equalization. Graph. Gems 1994, 4, 474–485. [Google Scholar]
  35. Cadena, L.; Zotin, A.; Cadena, F.; Korneeva, A.; Legalov, A.; Morales, B. Noise reduction techniques for processing of medical images. In Proceedings of the World Congress on Engineering, London, UK, 5–7 July 2017; pp. 5–9. [Google Scholar]
  36. Lestari, T.; Luthfi, A. Retinal blood vessel segmentation using Gaussian filter. J. Phys. Conf. Ser. 2019, 1376, 012023. [Google Scholar] [CrossRef]
  37. Cogan, T.; Cogan, M.; Tamil, L. MAPGI: Accurate identification of anatomical landmarks and diseased tissue in gastrointestinal tract using deep learning. Comput. Biol. Med. 2019, 111, 103351. [Google Scholar] [CrossRef] [PubMed]
  38. Kim, Y.W.; Krishna, A.V. A study on the effect of Canny edge detection on downscaled images. Pattern Recognit. Image Anal. 2020, 30, 372–381. [Google Scholar] [CrossRef]
  39. Welikala, R.A.; Remagnino, P.; Lim, J.H.; Chan, C.S.; Rajendran, S.; Kallarakkal, T.G.; Zain, R.B.; Jayasinghe, R.D.; Rimal, J.; Kerr, A.R. Fine-tuning deep learning architectures for early detection of oral cancer. In Proceedings of the Mathematical and Computational Oncology: Second International Symposium, ISMCO 2020, San Diego, CA, USA, 8–10 October 2020; Proceedings 2. Springer: Berlin/Heidelberg, Germany, 2020; pp. 25–31. [Google Scholar] [CrossRef]
  40. Sakr, A.S.; Soliman, N.F.; Al-Gaashani, M.S.; Pławiak, P.; Ateya, A.A.; Hammad, M. An efficient deep learning approach for colon cancer detection. Appl. Sci. 2022, 12, 8450. [Google Scholar] [CrossRef]
  41. Fuhrman, J.; Yip, R.; Zhu, Y.; Jirapatnakul, A.C.; Li, F.; Henschke, C.I.; Yankelevitz, D.F.; Giger, M.L. Evaluation of emphysema on thoracic low-dose CTs through attention-based multiple instance deep learning. Sci. Rep. 2023, 13, 1187. [Google Scholar] [CrossRef]
  42. Rolon-Mérette, D.; Ross, M.; Rolon-Mérette, T.; Church, K. Introduction to Anaconda and Python: Installation and setup. Quant. Methods Psychol. 2016, 16, S3–S11. [Google Scholar] [CrossRef]
  43. Haslwanter, T. An Introduction to Statistics with Python. In With Applications in the Life Sciences; Springer International Publishing: Cham, Switzerland, 2016. [Google Scholar]
  44. Gulli, A.; Pal, S. Deep Learning with Keras; Packt Publishing Ltd.: Birmingham, UK, 2017. [Google Scholar]
  45. Gollapudi, S. Learn Computer Vision Using OpenCV; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  46. Trappenberg, T.P. Fundamentals of Machine Learning; Oxford University Press: Oxford, UK, 2019. [Google Scholar]
  47. Wickham, H.; Chang, W.; Wickham, M.H. Package ‘ggplot2’. Creat. Elegant Data Vis. Using Gramm. Graph. Version 2016, 2, 1–189. [Google Scholar]
  48. Chollet, F. Deep Learning with Python; Simon and Schuster: New York, NY, USA, 2021. [Google Scholar]
  49. Goel, A.; Agarwal, A.; Vatsa, M.; Singh, R.; Ratha, N.K. DNDNet: Reconfiguring CNN for Adversarial Robustness. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 22–23. [Google Scholar]
  50. Zhang, J.; Lin, X.; Jiang, M.; Yu, Y.; Gong, C.; Zhang, W.; Tan, X.; Li, Y.; Ding, E.; Li, G. A Multi-Granularity Retrieval System for Natural Language-Based Vehicle Retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 3216–3225. [Google Scholar]
  51. Alaskar, H.; Hussain, A.; Al-Aseem, N.; Liatsis, P.; Al-Jumeily, D. Application of convolutional neural networks for automated ulcer detection in wireless capsule endoscopy images. Sensors 2019, 19, 1265. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Colon cancer in the wall of the colon. Figure created with Biorender.com (accessed on 1 March 2023).
Figure 1. Colon cancer in the wall of the colon. Figure created with Biorender.com (accessed on 1 March 2023).
Diagnostics 13 01721 g001
Figure 2. From left to right, the colon images of CT, colonoscopy, and histology, respectively, (a) before and (b) after pre-processing.
Figure 2. From left to right, the colon images of CT, colonoscopy, and histology, respectively, (a) before and (b) after pre-processing.
Diagnostics 13 01721 g002
Figure 3. An illustration of the DL approach used for the discrimination between normal colon (negative) and colon cancer (positive) to detect colon cancer.
Figure 3. An illustration of the DL approach used for the discrimination between normal colon (negative) and colon cancer (positive) to detect colon cancer.
Diagnostics 13 01721 g003
Figure 4. Transfer learning using feature extraction technique in developing the proposed CNN models.
Figure 4. Transfer learning using feature extraction technique in developing the proposed CNN models.
Diagnostics 13 01721 g004
Figure 5. The training performance results based on 5-fold cross-validation using ACC, PRE, REC, and F1 performance measures. Boxplot showing the performance measures of five DL models for CT, histology, and colonoscopy imaging modalities.
Figure 5. The training performance results based on 5-fold cross-validation using ACC, PRE, REC, and F1 performance measures. Boxplot showing the performance measures of five DL models for CT, histology, and colonoscopy imaging modalities.
Diagnostics 13 01721 g005
Figure 6. The test performance results based on 5-fold cross-validation using ACC, PRE, REC, and F1 performance measures. Boxplot showing the performance measures of five DL models for CT, histology, and colonoscopy imaging modalities.
Figure 6. The test performance results based on 5-fold cross-validation using ACC, PRE, REC, and F1 performance measures. Boxplot showing the performance measures of five DL models for CT, histology, and colonoscopy imaging modalities.
Diagnostics 13 01721 g006
Figure 7. The combined confusion matrices of two categories (normal colon (NEG) and colon cancer (POS)) based on five DL models using five-fold cross-validation on the test sets for each of the CT, histology, and colonoscopy.
Figure 7. The combined confusion matrices of two categories (normal colon (NEG) and colon cancer (POS)) based on five DL models using five-fold cross-validation on the test sets for each of the CT, histology, and colonoscopy.
Diagnostics 13 01721 g007
Figure 8. ROC curves for DL models applied to image datasets derived from the three studied imaging modalities using five-fold cross-validation. ROC: receiver operating characteristics.
Figure 8. ROC curves for DL models applied to image datasets derived from the three studied imaging modalities using five-fold cross-validation. ROC: receiver operating characteristics.
Diagnostics 13 01721 g008
Table 1. The number of CT, histology, and colonoscopy images used for detecting colon cancer.
Table 1. The number of CT, histology, and colonoscopy images used for detecting colon cancer.
DatasetModalityDistribution
POS 1NEG 2
TCGA-COADCT900-
CT COLONOGRAPHYCT-900
NCT-CRC-HE-100KNONORMHistology900900
HyperKvasirColonoscopy900900
1 POS refers to positive samples (cancer). 2 NEG refers to negative samples (normal).
Table 2. A performance comparison between the CT, histology, and colonoscopy imaging modalities using different deep learning (DL) models during the 5-fold cross-validation on test sets for accuracy (ACC), precision (PRE), recall (REC), F1, and area under the ROC curve (AUC). MACC is mean accuracy, MPRE is mean precision, and MREC is mean recall. MF1 is mean f1. MAUC is mean AUC. Bold represents the highest mean performance measure.
Table 2. A performance comparison between the CT, histology, and colonoscopy imaging modalities using different deep learning (DL) models during the 5-fold cross-validation on test sets for accuracy (ACC), precision (PRE), recall (REC), F1, and area under the ROC curve (AUC). MACC is mean accuracy, MPRE is mean precision, and MREC is mean recall. MF1 is mean f1. MAUC is mean AUC. Bold represents the highest mean performance measure.
Imaging ModalityMethodMACCMPREMREC MF1MAUC
CTVGG160.9700.9540.9880.9710.970
VGG190.9330.9350.9310.9330.933
ResNet152V20.9450.9390.9500.9450.945
MobileNetV20.8160.8890.7270.7980.816
DenseNet2010.9760.9610.9940.9770.976
HistologyVGG160.8680.8510.8920.8710.868
VGG190.8570.8730.8420.8550.857
ResNet152V20.8150.8190.8100.8130.815
MobileNetV20.6780.7140.6660.6750.678
DenseNet2010.9120.9110.9120.9120.912
ColonoscopyVGG160.9870.9970.9770.9870.987
VGG190.9820.9940.9710.9820.982
ResNet152V20.9640.9780.9500.9630.964
MobileNetV20.9450.9920.8950.9420.945
DenseNet2010.9910.9980.9840.9910.991
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alhazmi, W.; Turki, T. Applying Deep Transfer Learning to Assess the Impact of Imaging Modalities on Colon Cancer Detection. Diagnostics 2023, 13, 1721. https://doi.org/10.3390/diagnostics13101721

AMA Style

Alhazmi W, Turki T. Applying Deep Transfer Learning to Assess the Impact of Imaging Modalities on Colon Cancer Detection. Diagnostics. 2023; 13(10):1721. https://doi.org/10.3390/diagnostics13101721

Chicago/Turabian Style

Alhazmi, Wael, and Turki Turki. 2023. "Applying Deep Transfer Learning to Assess the Impact of Imaging Modalities on Colon Cancer Detection" Diagnostics 13, no. 10: 1721. https://doi.org/10.3390/diagnostics13101721

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop