Next Article in Journal
Relationship between Serum 25-Hydroxyvitamin D Levels and Hormonal Status in Infertile Women: A Retrospective Study
Next Article in Special Issue
Predictive Factors Correlated with Successful Early Endoscopic Removal of Pancreaticolithiasis in Chronic Pancreatitis after Extracorporeal Shock Wave Lithotripsy
Previous Article in Journal
Biology and Management of High-Grade Myxofibrosarcoma: State of the Art and Future Perspectives
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

The Advent of Domain Adaptation into Artificial Intelligence for Gastrointestinal Endoscopy and Medical Imaging

1
Division of Gastroenterology, Department of Internal Medicine, Dongguk University Ilsan Hospital, Dongguk University College of Medicine, Goyang 10326, Republic of Korea
2
Department of Intelligent Systems and Robotics, College of Electrical & Computer Engineering, Chungbuk National University, Cheongju 28644, Republic of Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2023, 13(19), 3023; https://doi.org/10.3390/diagnostics13193023
Submission received: 3 August 2023 / Revised: 1 September 2023 / Accepted: 12 September 2023 / Published: 22 September 2023
(This article belongs to the Special Issue Advances in the Diagnosis and Treatment of Hepatogastroenterology)

Abstract

:
Artificial intelligence (AI) is a subfield of computer science that aims to implement computer systems that perform tasks that generally require human learning, reasoning, and perceptual abilities. AI is widely used in the medical field. The interpretation of medical images requires considerable effort, time, and skill. AI-aided interpretations, such as automated abnormal lesion detection and image classification, are promising areas of AI. However, when images with different characteristics are extracted, depending on the manufacturer and imaging environment, a so-called domain shift problem occurs in which the developed AI has a poor versatility. Domain adaptation is used to address this problem. Domain adaptation is a tool that generates a newly converted image which is suitable for other domains. It has also shown promise in reducing the differences in appearance among the images collected from different devices. Domain adaptation is expected to improve the reading accuracy of AI for heterogeneous image distributions in gastrointestinal (GI) endoscopy and medical image analyses. In this paper, we review the history and basic characteristics of domain shift and domain adaptation. We also address their use in gastrointestinal endoscopy and the medical field more generally through published examples, perspectives, and future directions.

1. Introduction

Artificial intelligence (AI) has attracted significant attention in medical image analyses. Gastrointestinal (GI) endoscopy is an active area of AI research. Upper endoscopies, colonoscopies, and capsule endoscopies detect inflammation, bleeding foci, preneoplastic lesions, and gastrointestinal cancer. The goals of AI research are to improve our ability to detect abnormal lesions, as well as enhance gastrointestinal imaging and its quality control and clinical efficiency in real clinical environments. The applications of AI in GI endoscopy range from computer-aided detection (CAD) to objective assessments of the degree of bowel preparation. In particular, CAD can potentially reduce endoscopic reading time and dramatically increase accuracy. Convolutional neural networks (CNNs) are primary deep learning algorithms for endoscopic image processing. CNN-based algorithms have been successful in detecting a variety of esophageal, stomach, small bowel, and colorectal images using different imaging modalities [1,2,3,4,5,6].
At first glance, it appears that the implementation of an automated reading system for GI endoscopy should be a simple task. However, there are some major obstacles to it reaching the clinical level. The endoscopic images collected for the training and testing of the AI are not homogeneous. More specifically, there are various endoscopy manufacturers worldwide. In addition, even among products from the same manufacturer, the image characteristics differ depending on the filming environment or image-processing software. Therefore, the development of a universal algorithm that is also highly versatile is a challenge. When an AI algorithm trained with images obtained using a device from Company A is applied to images obtained using a device from Company B or another manufacturer, there is a concern that the accuracy of the AI may drop significantly. This is called a “domain shift problem”. It has already been reported that this domain shift problem occurs even when an endoscopic image from one hospital is applied to an AI module adapted in another hospital.
Domain adaptation is a promising tool for overcoming the problem of multimodal endoscopic image acquisition for AI development. It involves feature alignment, an image alignment process that typically uses a framework for image conversion to reduce the differences in appearance among images [7,8]. Domain adaptation technology forms the basis for creating a universal AI system for endoscopic image analyses.
This review encompasses the fundamentals and history of domain adaptation, its applications in GI endoscopy and the medical field more generally, and future perspectives.

2. Fundamentals and History of Domain Adaptation and Domain Shift

Domain adaptation began to be studied when generative adversarial networks (GAN), now one of the most widely used generative models, were introduced [9]. Previously, it was difficult to apply transformations between various domains, so it was difficult to generate a proper dataset. In a CycleGAN paper published in 2017, the concept of cycle consistency loss was introduced. This stated that the same image should be generated when converting from the source domain to the target domain, and then from the target domain to the source domain, even if there is no exact correspondence. Many studies on domain adaptation have now been conducted.
To implement domain adaptation, it is first necessary to understand deep-learning-based features, such as CNNs, which extract the important information from images. A GAN can be trained by understanding the different loss functions compared to those used in the existing CNN. It should be noted that the generator and discriminator are used simultaneously during training; however, only the generator is used to create an image for the new domain after learning. To implement this, the Python programming language is used, and open libraries for deep learning, such as TensorFlow or PyTorch, are installed [10,11,12].
GANs have had a significant impact on the field of AI by transforming supervised-learning-oriented deep learning systems into unsupervised learning. Previously, learning was possible only when all data were correct. However, the emergence of GANs, which can continuously generate new data without correct answers, has opened up new possibilities for AI; these include image generation and restoration, domain transformation, object detection, super resolution, and music generation. Domain adaptation refers to the adaptation of information from existing domains to new domains that are different but relevant. Here, the two main domains are divided into source and target domains and are assumed to have different dataset distributions (Figure 1) [7]. If the model learned from the source domain is to be applied to the target domain, a problem inevitably arises because the two domains have different distributions, which are usually referred to as “domain shifts”. Therefore, the goal of domain adaptation is to create a more robust model for these domains to prevent domain shifts.
Such a model can be leveraged to obtain a greater accuracy by increasing the amount of insufficient learning data, and existing learned networks can be exploited through domain translation without further learning. Because these domain conversion techniques can be applied to various existing object recognition problems, they have been used in a number of recently conducted studies [13]. GAN studies for domain transformation are divided into non-style code models such as Pix2Pix, CycleGAN, and DiscoGAN, and style code models such as MUNIT, DRIT, and StarGAN [14,15]. Depending on the structure of the model, the same data can produce different results. Therefore, it is important to use an appropriate network based on the learning data [16,17].
A generative adversarial network, or GAN, is a network in which constructors and identifiers compete with each other and generate data. The goal of a GAN is to generate data close to the distribution of the real data, and the generator attempts to generate fake data close to the real data so that the discriminator does not falsely discriminate (Figure 2). The goal of this process is to gradually improve the performances of the generator and discriminator and ultimately prevent the discriminator from distinguishing between real and fake data [18].
Pons et al. determined the performance before and after domain adaptation using a receiver operating characteristic curve (Figure 3) [19]. The performance of the baseline image before domain adaptation was extremely poor (area under the curve (AUC), 0.6488). After the domain adaptation with CycleGAN, the true positive rate of the image was significantly improved (AUC, 0.7341).

3. Application of Domain Adaptation GI Endoscopy and Medical Field

Table 1 summarizes the many studies that have been conducted on domain adaptation in the medical field.

3.1. Using Triplet Loss for Domain Adaptation in Wireless Capsule Endoscopy (WCE)

Since the announcement of the first WCE device in 2001, technological advances have progressed rapidly; new devices have been regularly introduced, offering increasingly good performances, better image resolution, better illumination, and larger fields of view. Today, WCE devices are produced by different manufacturers and offer a range of technical specifications [20,21].
We recently conducted a domain adaptation study using WCE. The latest AI models for automated WCE reading are useful with respect to reductions in reading time [22,23,24]. In addition, we used AI for automated abnormal lesion detection in small bowel WCE and found that the performance of the AI-assisted interpretation was comparable with that of experienced endoscopists for abnormal lesion detection without improvement. However, the performance of the optimized WCE AI algorithm deteriorated when it was applied to data from other hospitals. In addition, our previous study showed that a binary classification model (two categories of images: clinically insignificant images, including normal mucosa, bile, air bubbles, and debris; and significant lesion images, including inflammation, abnormal vascular lesions, and bleeding) produced excellent outcomes in an internal hospital test, with a high sensitivity even in unseen images. Unfortunately, this model showed subclinical outcomes in an external test at a third-party hospital. This phenomenon is also commonly experienced in AI applications to other images in real-world situations. This constitutes a major obstacle to the universal use of AI in medical fields. The domain shift problem has been also observed in WCE. AI engineers have adopted the domain adaptation technique to enable efficient algorithm learning by creating new virtual data and extracting the main features of a collected image. An A-image and a B-image are handled separately and divided into training and validation (test) sets. Two independent AI algorithms (A-AI and B-AI) are built using the training images from the two groups. These are trained to discriminate between normal and significantly abnormal images. After the training, each AI algorithm is applied to a homogeneous test set. These algorithms are then tested using a heterogeneous test set (e.g., applying A-AI to the B-image test set, or applying B-AI to the A-image test set) to determine whether the performance improves or deteriorates. Their performances are evaluated using heterogeneous test images after several up-to-date domain adaptation techniques are applied to the test images. Thus, we attempted to evaluate the extent to which the domain adaptation technique can improve the reading accuracy in heterogeneous images (Figure 4). We tested three different domain adaptation architectures, introduced from 2017 to present: CycleGAN, discoGAN, and MUNIT. These domain adaptation methods enable the original images of WCE to be transformed and reconstructed into a new image to match the distribution of the target domain. This process allows the newly created image to possess a style similar to the heterogenous capsule image. We also compared various domain adaptation techniques for an EfficientNet-based algorithm and ResNet-based algorithm. We found that the domain adaptation was effective regardless of the algorithm used. In addition, domain adaptation using CycleGAN resulted in an additional AI performance improvement. A similar improvement in terms of performance improvement was obtained using discoGAN. The improvement in terms of efficacy outcome was the greatest when the domain adaptation was performed with CycleGAN.
It is not surprising that, if a model is trained with data from an older capsule, it may not yield the expected results when it is evaluated with a newer capsule, because the same distribution of data is not guaranteed. However, it is inefficient and expensive to abandon an old database and create a new one from scratch each time a new device is developed. To overcome this problem, the authors propose a domain adaptation method based on deep metric learning using triplet loss. The aim of this method is to adapt an embedding space trained with a large training dataset to a new domain, in which comparatively few labeled images are available. The embedding space is adapted by generating triplets of images from both domains with the goal of two images in the same category being closer than images belonging to different domains. The research results show that, by using a small, labeled dataset from the new domain, the embedding space can perform well. Effective results may be readily obtained in the new environment using domain adaptation with only a few labeled images from older camera systems [25].

3.2. Colonoscopy Polyp Detection: Domain Adaptation from Medical Report Images to Real-Time Videos

Manually annotating polyp regions in large-scale video datasets is time-consuming and expensive, and has thus limited the development of deep learning techniques. To compensate for this, researchers have used labeled images to train target models and infer colonoscopy videos. However, there are many problems with image-based training and video-based inference, including domain differences, a lack of positive samples, and temporal smoothness. To address these issues, the use of an image-video-joint polyp detection network (Ivy-Net), which is a type of domain adaptation, has been proposed to address the domain gap between colonoscopy images from historical medical reports and real-time videos. In Ivy-Net, a modified mix-up is utilized to generate training data by combining positive and negative video frames at the pixel level, which can learn domain-adaptive representations and augment positive samples. Experiments on collected datasets have demonstrated that Ivy-Net achieves state-of-the-art results and significantly improves the average precision of the polyp detection in colonoscopy videos [26,27,28].

3.3. Unsupervised Adversarial Domain Adaptation for Barrett’s Segmentation

Because Barrett’s esophagus (BE) is a precancerous lesion, it is important to accurately identify the BE region, so that patients may be adequately monitored and minimally invasive therapy may be administered. Automated segmentation using AI helps clinical endoscopists to evaluate this BE area more accurately and thus determine a range of treatments [29,30]. The existing automated segmentation methods use CNN-based supervised models [31]. These supervised models require a large number of manual annotations that incorporate all data variability into the AI training data, and these fully supervised models often cannot be generalized to different imaging modalities because of domain shifts. This problem can be alleviated by applying unsupervised domain adaptation (UDA). UDA is trained on white-light images as the source domain, and is well-adapted for generalization to produce segmentation on different imaging modalities as the target domain; these include narrow-band imaging (NBI) and post-acetic-acid (PAA) white-light imaging. This approach has been found to provide a generalized prediction of the segmentation masks of unlabeled endoscopy images in cross-modalities and improve performance with respect to both NBI and PAA images. This method does not rely on the existence of target labels and provides accurate and stable results under different imaging conditions. Experimental results have demonstrated the greater effectiveness of UDA-based models compared to traditional supervised models in reconstructing the BE area, an early cancer precursor [32].

3.4. Domain Adaptation for Alzheimer’s Disease Classification

Magnetic resonance imaging is an excellent diagnostic technique. When using computer-aided diagnosis to diagnose dementia, characterizing of the brain anatomy is promising for diagnosing and classifying Alzheimer’s disease (AD), mild cognitive impairment, and normal controls. Large, multicenter datasets are available for studying AD and supporting the training of complex classification models. Interestingly, one study using this classification model found that all the participating groups overestimated the accuracy of their method. One of the main reasons for this reduced classification accuracy was the variation in the distribution between the training and test data. Such problems involve domain adaptation, in which the model is taught using a source dataset and transferred to a target dataset with different properties based on instance weighting. They used supervised domain adaptation, in which the source domain was the training domain with labeled data, whereas the target domain was the test domain with only a fraction of labeled data. This could be a possible AD classification method for improving recognition rate, regardless of the data source [33,34,35].

3.5. Semi-Supervised Learning with GANs for Chest X-ray Classification with the Ability of Data Domain Adaptation

Because of privacy laws, medical industry standards, and a lack of integration into medical information systems, the sources of medical imaging data are not as rich as those of other fields in computer vision. Therefore, it is a challenge to develop deep learning algorithms for medical imaging. Even if data are available, unstructured or inadequate labeling becomes an obstacle to utilizing these data. To address this problem, medical images may be annotated; however, this is a time-consuming and costly process. As a result, when deep-learned classifiers are trained on a particular training dataset and then tested in production on data from a different domain source, their performance and accuracy are determined. Madani et al. addressed the problems of labeled data scarcity and data domain differences using GANs. They confirmed that deep GANs can learn the visual structures of medical imaging domain sources (particularly chest X-rays). They proposed a semi-supervised architecture for GANs, which is capable of learning from both labeled and unlabeled images. Their results showed that, when labeled data are limited, a semi-supervised GAN-based network requires one-order-of-magnitude-less labeled training data to achieve a performance comparable with that of a supervised CNN classifier. In other words, a performance similar to that of supervised training techniques is achieved with a considerably reduced annotation effort. They attributed this result to GANs being able to learn the structures in unlabeled data when using unsupervised learning methods, which significantly offsets the low number of labeled data samples [36,37,38].

3.6. UDA-Based Coronavirus Disease of 2019 (COVID-19) Infection Segmentation Network

The automatic segmentation of infected lung regions in computed tomography (CT) images has proven to be an effective diagnostic tool for COVID-19. However, because of the limited number of pixel-level labeled medical images, accurate segmentation remains a major challenge. More recently, authors generated synthetic COVID-19 CT data to promote the computer-aided diagnostic ability for COVID-19, making it possible to train deep models on synthetic images and computer-generated annotations. However, this study also found that a model directly trained on synthetic data may fail to produce accurate results for real COVID-19 CT images, because of domain shift. To resolve this problem, the authors proposed a UDA-based segmentation network to improve the segmentation performance for the infected areas in the COVID-19 CT images. They proposed making full use of synthetic data and limited unlabeled real COVID-19 CT images to train the segmentation network jointly in order to introduce a richer diversity. This approach reduced domain shift by forcing the features from different domains to fool the discriminator, leading to features from different domains exhibiting a similar distribution. This method played an important role in diagnosing COVID-19 by quantifying the infected areas of the lungs. Domain adaptation also demonstrated positive effects on medical image segmentation [39,40].
Table 1. Example of studies on application of domain adaptation in the medical images analysis.
Table 1. Example of studies on application of domain adaptation in the medical images analysis.
ReferenceMedical InstrumentTaskModuleResult
Laiz et al. (2019) [25]Capsule endoscopyImprove the generalization of a model over different datasets from different versions of WCE hardware.Deep metric learning, based on the triplet loss functionJust a few labeled images from a newer camera set, a model that has been trained with images from older systems can be easily adapted to the new environment.
Zhan et al. (2020) [28]ColonoscopyColon polyp detection, images to real-time videosIvy-NetIvy-Net to alleviate the domain gap between colonoscopy images from historical medical reports and real-time videos.
Celik et al. (2012) [32]GastroscopyBarret’s esophagus area segmentationUDAUDA method generalizes on different imaging modalities showing improved segmentation accuracy.
Wachinger et al. (2016) [35]MRIAlzheimer’s disease classificationSupervised domain adaptation(SDA)Domain adaptation with instance weighting yields the best classification results
Madani et al. (2018) [38]Chest X-rayAbnormality detectionGANsAnnotation effort is reduced to achieve similar performance through supervised training techniques.
Chen et al. (2021) [39]CTAutomatic segmentation of infection areaUDASegmentation network to learn the domain-invariant feature, so that the robust feature can be used for segmentation.

4. Perspective and Future Direction

The latest AI models for automated GI endoscopy reading have fast learning speeds and fairly high reading accuracies. Thus, they have a high potential for clinical use. However, these AI models have several shortcomings and exhibit significant performance degradation when the data format or domain is slightly different from the training data format. This issue is called a “domain shift problem”. This is one of the major factors preventing the use of AIs at the preclinical level. This domain shift problem has been observed in GI endoscopy and other medical imaging processes, because gathered medical images are obtained using different scanners with different scanning parameters and involve different subject cohorts. As a result, and as mentioned above, heterogeneity among medical image datasets is inevitable.
For this reason, domain adaptation techniques have gained attention among developers. A domain is related to a specific dataset’s feature space and features a marginal probability distribution [41]. Originally, AI engineers have adopted domain adaptation techniques to enable efficient algorithm learning by creating new virtual data and extracting the main features of obtained images. However, image shifts through domain adaptation can minimize the distribution gap among different but partially related domains in medical image analyses. For domain adaptation, it is assumed that the domain feature spaces and tasks remain the same, but there is a difference in the marginal distributions between the source and target domains. Many applications of domain adaptation in GI endoscopy and medical image analyses have been reported. Laiz et al. first proposed the concept of triplet loss function for domain adaptation in capsule endoscopy, and reported that it could improve the reading accuracy by improving the generalization of the datasets obtained from different systems [25]. Zhan et al. improved the polyp detection in colonoscopy videos by using Ivy-Net to bridge the domain gap between colonoscopy images and real-time videos [28]. Celik et al. applied a UDA framework for the segmentation of BE, which is a precancerous lesion [32]. Wachinger et al. used AD classification to improve the recognition rate, regardless of the data source, using supervised domain adaptation, in which the source domain was the training domain with labeled data [35]. Madani et al. confirmed the positive effects of domain adaptation on chest X-ray classification [38]. Chen et al. developed a novel domain adaptation module that could achieve a good segmentation performance on COVID-19 CT images, even when no annotations were provided [39].
Many researchers have used only supervised learning for machine learning strategies in AI studies on medical images. Active learning and unsupervised learning are still not available because of their low outcome accuracy. It takes a lot of time to collect, classify, and label the learning materials for an AI algorithm’s development. If domain adaptation could be applied to new medical images, the previously developed algorithm would be able to bring about a robust accuracy to those images. Domain adaptation could be also applied to medical images obtained using old medical devices. Of course, it is very useful when AI learning materials are lacking.
Can AI-assisted interpretation replace the current time-consuming conventional reading in real clinical fields? In terms of the immediate future, our answer is no. It may be some time before such a situation comes to pass. Although many published studies have shown the very high sensitivity and specificity of AI algorithms in the interpretation of medical images, it should be noted that the training and testing sets in these studies were designed for selected images, not medical images obtained from real-world clinical environments.
Actual medical images are images in which various lesions, non-specific findings, normal variants, blurring, and images obtained from different hospitals are mixed. These are different from the testing sets that are used by researchers. The reason for the low accuracy of AI is that too many false positives are produced in real-world applications. Although the test results are usually very good in limited conditions, when compared to external results from other institutes, a low accuracy is typically reported, indicating suboptimal results. The problem of overfitting is important in this regard. To overcome the overfitting problem, domain adaptation is an engineering technique that helps AI to be actually and universally used in real clinical fields. We would like to emphasize that the availability of good data is still the first most important point in AI studies. Large amounts of data are important, as is an even distribution of the various different images constituting the training data set. For good data, a reference standard should be established by consensus from a panel of experienced professionals.
Domain adaptation has the potential to be utilized in various fields of medical image analysis. In addition, the need to label a large-scale dataset can be alleviated by the use of a virtually generated image using domain adaptation. By such means, a time-consuming, laborious, and expensive task can be avoided. Moreover, this technique is expected to improve the accuracy of AI reading, while solving the problem of heterogeneous image distribution in multicenter studies. Ultimately, the AI model developed for lesion detection in GI endoscopy and other medical fields can easily be utilized by domain adaptation.

5. Conclusions

Automated lesion detection through AI can be seen as a promising field of research. However, as images with different characteristics are extracted, depending on the device manufacturer and the imaging environment, the so-called “domain shift problem” occurs, in which the developed AI has a poor versatility. Domain adaptation is a tool that can generate a new, converted image suitable for other domains. This is a promising tool for reducing the appearance differences among images collected from different devices and environments. It is expected to improve the AI reading accuracy for heterogenous image distribution in medical image analyses.

Author Contributions

Conception and design: Y.J.L. and Y.B.H. Drafting of the article: M.J.K., S.H.K., S.M.K. and J.H.N. Critical revision: Y.J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by a grant (HI19C0665) from the Korean Health Technology R & D project through the Korean Health Industry Development Institute funded by the Ministry of Health and Welfare, Republic of Korea and Dongguk University Research Fund (2023) and the National Research Foundation of Korea (NRF) grant (2021R1G1A1094851) funded by the Korea government (MSIT).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AIartificial intelligence
AUCarea under the curve
CNNconvolutional neural network.
MRImagnetic resonance imaging
ROCreceiver operating characteristic
GIgastrointestinal
CADcomputer-aided detection
GANgenerative adversarial network
WCEwireless capsule endoscopy
BEBarrett’s esophagus
UDAunsupervised domain adaptation
SDAsupervised domain adaptation
NBInarrow-band imaging
PAApost-acetic acid
WLwhite light
ADAlzheimer’s disease
COVID-19Coronavirus disease of 2019
CTcomputed tomography

References

  1. Sumiyama, K.; Futakuchi, T.; Kamba, S.; Matsui, H.; Tamai, N. Artificial intelligence in endoscopy: Present and future perspectives. Dig. Endosc. 2021, 33, 218–230. [Google Scholar] [CrossRef]
  2. Nam, J.H.; Lee, K.H.; Lim, Y.J. Examination of Entire Gastrointestinal Tract: A Perspective of Mouth to Anus (M2A) Capsule Endoscopy. Diagnostics 2021, 11, 1367. [Google Scholar] [CrossRef]
  3. Sadagopan, R.; Ravi, S.; Adithya, S.V.; Vivekanandhan, S. PolyEffNetV1: A CNN based colorectal polyp detection in colonoscopy images. Proc. Inst. Mech. Eng. H 2023, 237, 406–418. [Google Scholar] [CrossRef] [PubMed]
  4. Ma, H.; Wang, L.; Chen, Y.; Tian, L. Convolutional neural network-based artificial intelligence for the diagnosis of early esophageal cancer based on endoscopic images: A meta-analysis. Saudi J. Gastroenterol. 2022, 28, 332–340. [Google Scholar] [CrossRef] [PubMed]
  5. Vu, H.; Manh, X.H.; Duc, B.Q.; Ha, V.K.; Dao, V.H.; Nguyen, P.B.; Hoang, B.L.; Vu, T.H. Labelling stomach anatomical locations in upper gastrointestinal endoscopic images using a cnn. In Proceedings of the 10th International Symposium on Information and Communication Technology, Ha Long Bay, Vietnam, 4–6 December 2019; pp. 362–369. [Google Scholar]
  6. Kim, S.H.; Hwang, Y.; Oh, D.J.; Nam, J.H.; Kim, K.B.; Park, J.; Song, H.J.; Lim, Y.J. Efficacy of a comprehensive binary classification model using a deep convolutional neural network for wireless capsule endoscopy. Sci. Rep. 2021, 11, 17479. [Google Scholar] [CrossRef]
  7. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  8. Choudhary, A.; Tong, L.; Zhu, Y.; Wang, M.D. Advancing medical imaging informatics by deep learning-based domain adaptation. Yearb. Med. Inform. 2020, 29, 129–138. [Google Scholar] [CrossRef]
  9. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  10. Rampasek, L.; Goldenberg, A. TensorFlow: Biology’s Gateway to Deep Learning? Cell Syst. 2016, 2, 12–14. [Google Scholar] [CrossRef] [PubMed]
  11. Ziller, A.; Usynin, D.; Braren, R.; Makowski, M.; Rueckert, D.; Kaissis, G. Medical imaging deep learning with differential privacy. Sci. Rep. 2021, 11, 13524. [Google Scholar] [CrossRef]
  12. Mishra, P. CNN and RNN Using PyTorch. In PyTorch Recipes: A Problem-Solution Approach to Build, Train and Deploy Neural Network Models; Springer: Berlin/Heidelberg, Germany, 2022; pp. 49–115. [Google Scholar]
  13. Huang, X.; Liu, M.-Y.; Belongie, S.; Kautz, J. Multimodal unsupervised image-to-image translation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 172–189. [Google Scholar]
  14. Kim, T.; Cha, M.; Kim, H.; Lee, J.K.; Kim, J. Learning to discover cross-domain relations with generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 1857–1865. [Google Scholar]
  15. Choi, Y.; Choi, M.; Kim, M.; Ha, J.-W.; Kim, S.; Choo, J. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8789–8797. [Google Scholar]
  16. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  17. Lee, H.-Y.; Tseng, H.-Y.; Huang, J.-B.; Singh, M.; Yang, M.-H. Diverse image-to-image translation via disentangled representations. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 35–51. [Google Scholar]
  18. Baek, K.; Choi, Y.; Uh, Y.; Yoo, J.; Shim, H. Rethinking the truly unsupervised image-to-image translation. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop, Montreal, BC, Canada, 11–17 October 2021; pp. 14154–14163. [Google Scholar]
  19. Pons, G.; El Ali, A.; Cesar, P. ET-CycleGAN: Generating thermal images from images in the visible spectrum for facial emotion recognition. In Proceedings of the Companion Publication of the 2020 International Conference on Multimodal Interaction, Virtual Event, The Netherlands, 25–29 October 2020; pp. 87–91. [Google Scholar]
  20. Sushma, B.; Aparna, P. Recent developments in wireless capsule endoscopy imaging: Compression and summarization techniques. Comput. Biol. Med. 2022, 149, 106087. [Google Scholar] [CrossRef]
  21. Muhammad, K.; Khan, S.; Kumar, N.; Del Ser, J.; Mirjalili, S. Vision-based personalized wireless capsule endoscopy for smart healthcare: Taxonomy, literature review, opportunities and challenges. Future Gener. Comput. Syst. 2020, 113, 266–280. [Google Scholar] [CrossRef]
  22. Aoki, T.; Yamada, A.; Aoyama, K.; Saito, H.; Fujisawa, G.; Odawara, N.; Kondo, R.; Tsuboi, A.; Ishibashi, R.; Nakada, A.; et al. Clinical usefulness of a deep learning-based system as the first screening on small-bowel capsule endoscopy reading. Dig. Endosc. 2020, 32, 585–591. [Google Scholar] [CrossRef] [PubMed]
  23. Kim, S.H.; Lim, Y.J. Artificial Intelligence in Capsule Endoscopy: A Practical Guide to Its Past and Future Challenges. Diagnostics 2021, 11, 1722. [Google Scholar] [CrossRef] [PubMed]
  24. Oh, D.J.; Hwang, Y.; Lim, Y.J. A Current and Newly Proposed Artificial Intelligence Algorithm for Reading Small Bowel Capsule Endoscopy. Diagnostics 2021, 11, 1183. [Google Scholar] [CrossRef]
  25. Laiz, P.; Vitria, J.; Seguí, S. Using the triplet loss for domain adaptation in WCE. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  26. Chen, P.J.; Lin, M.C.; Lai, M.J.; Lin, J.C.; Lu, H.H.; Tseng, V.S. Accurate Classification of Diminutive Colorectal Polyps Using Computer-Aided Analysis. Gastroenterology 2018, 154, 568–575. [Google Scholar] [CrossRef] [PubMed]
  27. Kalogeiton, V.; Ferrari, V.; Schmid, C. Analysing Domain Shift Factors between Videos and Images for Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 2327–2334. [Google Scholar] [CrossRef]
  28. Zhan, Z.-Q.; Fu, H.; Yang, Y.-Y.; Chen, J.; Liu, J.; Jiang, Y.-G. Colonoscopy polyp detection: Domain adaptation from medical report images to real-time videos. arXiv 2020, arXiv:2012.15531. [Google Scholar]
  29. Hamade, N.; Sharma, P. ‘Artificial intelligence in Barrett’s Esophagus’. Ther. Adv. Gastrointest. Endosc. 2021, 14, 26317745211049964. [Google Scholar] [CrossRef]
  30. Dumoulin, F.L.; Rodriguez-Monaco, F.D.; Ebigbo, A.; Steinbruck, I. Artificial Intelligence in the Management of Barrett’s Esophagus and Early Esophageal Adenocarcinoma. Cancers 2022, 14, 1918. [Google Scholar] [CrossRef]
  31. Ohmori, M.; Ishihara, R.; Aoyama, K.; Nakagawa, K.; Iwagami, H.; Matsuura, N.; Shichijo, S.; Yamamoto, K.; Nagaike, K.; Nakahara, M.; et al. Endoscopic detection and differentiation of esophageal lesions using a deep neural network. Gastrointest. Endosc. 2020, 91, 301–309.e1. [Google Scholar] [CrossRef]
  32. Celik, N.; Gupta, S.; Ali, S.; Rittscher, J. Unsupervised Adversarial Domain Adaptation For Barrett’s Segmentation. arXiv 2020, arXiv:2012.05316. [Google Scholar]
  33. Orbes-Arteaga, M.; Varsavsky, T.; Sudre, C.H.; Eaton-Rosen, Z.; Haddow, L.J.; Sorensen, L.; Nielsen, M.; Pai, A.; Ourselin, S.; Modat, M.; et al. Multi-domain Adaptation in Brain MRI Through Paired Consistency and Adversarial Learning. Domain Adapt. Represent. Transf. Med. Image Learn. Less Labels Imperfect Data 2019, 2019, 54–62. [Google Scholar] [CrossRef]
  34. Yu, M.; Guan, H.; Fang, Y.; Yue, L.; Liu, M. Domain-Prior-Induced Structural MRI Adaptation for Clinical Progression Prediction of Subjective Cognitive Decline. Med. Image Comput. Comput. Assist. Interv. 2022, 13431, 24–33. [Google Scholar] [CrossRef] [PubMed]
  35. Wachinger, C.; Reuter, M.; Alzheimer’s Disease Neuroimaging, I.; Australian Imaging, B.; Lifestyle flagship study of, a. Domain adaptation for Alzheimer’s disease diagnostics. Neuroimage 2016, 139, 470–479. [Google Scholar] [CrossRef]
  36. Çallı, E.; Sogancioglu, E.; van Ginneken, B.; van Leeuwen, K.G.; Murphy, K. Deep learning for chest X-ray analysis: A survey. Med. Image Anal. 2021, 72, 102125. [Google Scholar] [CrossRef] [PubMed]
  37. Guan, H.; Liu, M. Domain Adaptation for Medical Image Analysis: A Survey. IEEE Trans. Biomed. Eng. 2022, 69, 1173–1185. [Google Scholar] [CrossRef]
  38. Madani, A.; Moradi, M.; Karargyris, A.; Syeda-Mahmood, T. Semi-supervised learning with generative adversarial networks for chest X-ray classification with ability of data domain adaptation. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 1038–1042. [Google Scholar]
  39. Chen, H.; Jiang, Y.; Loew, M.; Ko, H. Unsupervised domain adaptation based COVID-19 CT infection segmentation network. Appl. Intell. 2022, 52, 6340–6353. [Google Scholar] [CrossRef]
  40. Xu, G.X.; Liu, C.; Liu, J.; Ding, Z.; Shi, F.; Guo, M.; Zhao, W.; Li, X.; Wei, Y.; Gao, Y.; et al. Cross-Site Severity Assessment of COVID-19 From CT Images via Domain Adaptation. IEEE Trans. Med. Imaging 2022, 41, 88–102. [Google Scholar] [CrossRef]
  41. Feuz, K.D.; Cook, D.J. Transfer Learning across Feature-Rich Heterogeneous Feature Spaces via Feature-Space Remapping (FSR). ACM Trans. Intell. Syst. Technol. 2015, 6, 1–27. [Google Scholar] [CrossRef]
Figure 1. Occurrence of domain shift and process of domain adaptation.
Figure 1. Occurrence of domain shift and process of domain adaptation.
Diagnostics 13 03023 g001
Figure 2. Principle of generative adversarial network (GAN).
Figure 2. Principle of generative adversarial network (GAN).
Diagnostics 13 03023 g002
Figure 3. Changes in the receiver operating characteristic curves of artificial intelligence after domain adaptation.
Figure 3. Changes in the receiver operating characteristic curves of artificial intelligence after domain adaptation.
Diagnostics 13 03023 g003
Figure 4. Example: application of domain adaptation in capsule endoscopy.
Figure 4. Example: application of domain adaptation in capsule endoscopy.
Diagnostics 13 03023 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, M.J.; Kim, S.H.; Kim, S.M.; Nam, J.H.; Hwang, Y.B.; Lim, Y.J. The Advent of Domain Adaptation into Artificial Intelligence for Gastrointestinal Endoscopy and Medical Imaging. Diagnostics 2023, 13, 3023. https://doi.org/10.3390/diagnostics13193023

AMA Style

Kim MJ, Kim SH, Kim SM, Nam JH, Hwang YB, Lim YJ. The Advent of Domain Adaptation into Artificial Intelligence for Gastrointestinal Endoscopy and Medical Imaging. Diagnostics. 2023; 13(19):3023. https://doi.org/10.3390/diagnostics13193023

Chicago/Turabian Style

Kim, Min Ji, Sang Hoon Kim, Suk Min Kim, Ji Hyung Nam, Young Bae Hwang, and Yun Jeong Lim. 2023. "The Advent of Domain Adaptation into Artificial Intelligence for Gastrointestinal Endoscopy and Medical Imaging" Diagnostics 13, no. 19: 3023. https://doi.org/10.3390/diagnostics13193023

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop