Next Article in Journal
A Data Analytic Monitoring with IoT System of the Reproductive Conditions of the Red Worm as a Product Diversification Strategy
Previous Article in Journal
Seismic Elastic Parameter Inversion via a FCRN and GRU Hybrid Network with Multi-Task Learning
Previous Article in Special Issue
Improved Stress Classification Using Automatic Feature Selection from Heart Rate and Respiratory Rate Time Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Applying Deep Learning to Medical Imaging: A Review

1
National Key Laboratory of Antennas and Microwave Technology, Xidian University, Xi’an 710071, China
2
Department of Industrial Engineering, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2023, 13(18), 10521; https://doi.org/10.3390/app131810521
Submission received: 30 April 2023 / Revised: 20 May 2023 / Accepted: 8 September 2023 / Published: 21 September 2023
(This article belongs to the Special Issue Medical Imaging Using Machine Learning and Deep Learning)

Abstract

:
Deep learning (DL) has made significant strides in medical imaging. This review article presents an in-depth analysis of DL applications in medical imaging, focusing on the challenges, methods, and future perspectives. We discuss the impact of DL on the diagnosis and treatment of diseases and how it has revolutionized the medical imaging field. Furthermore, we examine the most recent DL techniques, such as convolutional neural networks (CNNs), recurrent neural networks (RNNs), and generative adversarial networks (GANs), and their applications in medical imaging. Lastly, we provide insights into the future of DL in medical imaging, highlighting its potential advancements and challenges.

1. Introduction

1.1. Background and Motivation

Medical imaging has been a critical component of modern healthcare, providing clinicians with vital information for the diagnosis, treatment, and monitoring of various diseases [1]. Traditional image analysis techniques often rely on handcrafted features and expert knowledge, which can be time-consuming and subject to human error [2]. In recent years, machine learning (ML) methods have been increasingly applied to medical image analysis to improve efficiency and reduce potential human errors. These methods, including Support Vector Machines (SVMs), decision trees, random forests, and logistic regression, have shown success in tasks such as image segmentation, object detection, and disease classification. These ML methods typically involve the manual selection and extraction of features from the medical images, which are then used for prediction or classification. With the rapid development of deep learning (DL) technologies, there has been a significant shift toward leveraging these powerful tools to improve the accuracy and efficiency of medical image analysis [3]. Unlike traditional ML methods, DL models are capable of automatically learning and extracting hierarchical features from raw data. Deep learning, a subfield of machine learning (ML), has made remarkable advancements in recent years, particularly in image recognition and natural language processing tasks [4]. This success is primarily attributed to the development of artificial neural networks (ANN) with multiple hidden layers, which allow for the automatic extraction and learning of hierarchical features from raw data [5]. Consequently, DL techniques and network-based computation have been widely adopted in various applications, including autonomous driving, robotics, natural language understanding [6], and a large number of engineering computation cases [7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43].
In the medical imaging domain, DL has shown great potential for enhancing the quality of care and improving patient outcomes [44]. By automating the analysis of medical images, DL algorithms can aid in the early detection of diseases, streamline clinical workflows, and reduce the burden on healthcare professionals [45]. In addition, DL also plays a significant role in the credibility verification of reported medical data. For instance, it can be utilized to identify anomalies or inconsistencies in the data, thereby ensuring the reliability of the data used for diagnosis or treatment planning. DL models can also help in validating the authenticity of medical images, which is crucial in today’s digital age where data manipulation has become increasingly sophisticated. Moreover, DL models can be trained to predict disease progression and treatment response, thereby contributing to personalized medicine and the optimization of therapeutic strategies [46].
In our study, we specifically discuss the potential of DL models in medical imaging. We have discovered that deep learning techniques have been revolutionizing the medical imaging research. These findings underline the potential of DL techniques to further advance the field of medical imaging, opening new avenues for diagnosis and treatment strategies. This paper details these methods, results, and the implications of these findings for future research.

1.2. DL Techniques

Several DL techniques have been applied to medical imaging [47,48,49,50,51,52], with convolutional neural networks (CNNs) being the most prevalent [53]. CNNs are particularly suited for image analysis tasks due to their ability to capture local spatial patterns and automatically learn hierarchical representations from input images [54]. Other DL techniques that have been applied to medical imaging include recurrent neural networks (RNNs), which are well-suited for handling sequential data, and generative adversarial networks (GANs), which can generate new samples from learned data distributions [55]. In assessing the performance of our DL models in medical image diagnosis, several evaluation metrics are commonly employed, including Receiver Operating Characteristic (ROC) curves and confusion matrices, among other techniques [1,2,3]. The ROC curve is a graphical plot that illustrates the diagnostic ability of our DL models as its discrimination threshold is varied. It presents the trade-off between sensitivity (or True Positive Rate) and specificity (1–False Positive Rate), providing a measure of how well our models distinguish between classes. The Area Under the ROC Curve (AUC) is also considered, which provides a single metric to compare model performance. On the other hand, confusion matrices provide a summary of prediction results on a classification problem. The number of correct and incorrect predictions is counted and broken down by each class. This offers a more granular view of the model performance, including metrics such as precision, recall, and F1-score, which are crucial when dealing with imbalanced classes.

1.3. Medical Imaging Modalities

There are various medical imaging modalities used in clinical practice, each providing unique information and serving specific diagnostic purposes [56]. Some of the most common modalities include magnetic resonance imaging (MRI), computed tomography (CT), positron emission tomography (PET), ultrasound imaging, and optical coherence tomography (OCT) [57], as shown in Figure 1. DL techniques have been successfully applied to these modalities for tasks such as image segmentation, classification, reconstruction, and registration [46].

1.4. Challenges and Opportunities

Despite the promising results achieved by DL in medical imaging, several challenges remain [47,48,49,50,51,52]. One major challenge is the limited availability of annotated medical image datasets due to the time-consuming and costly nature of manual annotations [58]. Additionally, data privacy concerns and the sharing of sensitive patient information pose significant obstacles to the development of large-scale, multi-institutional datasets [59]. Another challenge is the interpretability of DL models, as they often act as “black boxes” that provide limited insights into their decision-making processes [60]. Ensuring the explainability and trustworthiness of these models is crucial for their adoption in clinical practice, as clinicians need to understand the rationale behind their predictions [61].
Despite these challenges, DL in medical imaging presents numerous opportunities for advancing healthcare and improving patient outcomes. With ongoing research, interdisciplinary collaboration, and the development of more sophisticated algorithms, DL has the potential to revolutionize medical imaging and contribute significantly to the future of medicine.

2. Deep Learning Techniques in Medical Imaging

Deep learning techniques in medical imaging can serve a wide array of functions, both in terms of the acquisition of medical images and the identification of pathologies within these images. Specifically, these techniques are leveraged not only to enhance the quality of images obtained through various modalities but also to enable effective and efficient identification of pathological markers within these images. For example, convolutional neural networks (CNNs) can be used in the reconstruction of images from MRI scanners, enhancing the resolution of the obtained images and thereby allowing for a clearer visualization of potential pathologies [53]. Moreover, CNNs are particularly adept at analyzing these images postacquisition, identifying key features within these images that could point toward specific pathologies [54]. This dual functionality—improving the acquisition of images and aiding in the identification of pathologies—is a key strength of deep learning techniques in the field of medical imaging. Throughout this section, we will discuss three major types of deep learning techniques used in medical imaging: convolutional neural networks (CNNs), recurrent neural networks (RNNs), and generative adversarial networks (GANs). For each technique, we will detail its basic concepts, architecture and applications, role in image acquisition, and pathology detection, along with transfer learning approaches and the limitations and challenges faced.

2.1. Convolutional Neural Networks (CNNs)

2.1.1. Basic Concepts

Convolutional neural networks (CNNs) are a class of DL models designed specifically for image analysis tasks [4]. Its basic mechanism has been indicated in Figure 2. CNNs consist of multiple layers, including convolutional, pooling, and fully connected layers, which work together to learn hierarchical representations of input images [62]. Convolutional layers are responsible for extracting local features from images, such as edges, corners, and textures, while pooling layers help reduce the spatial dimensions of feature maps, improving computational efficiency and reducing overfitting [63]. Finally, fully connected layers enable the integration of local features into global patterns, enabling the network to perform image classification or other desired tasks [6].

2.1.2. Architectures and Applications

Several CNN architectures have been proposed and widely adopted in medical imaging applications [64,65,66]. Some of the most notable architectures include LeNet [67], AlexNet [63], VGGNet [62], ResNet [53], and DenseNet [68]. These architectures have been applied to various medical imaging tasks, such as image segmentation, classification, detection, and registration [69,70].

2.1.3. Transfer Learning

Transfer learning is a popular approach in DL, where a pretrained model is fine-tuned for a new task or domain, leveraging the knowledge acquired during the initial training [71]. This technique is particularly useful in medical imaging [72,73,74], where annotated datasets are often limited in size [75]. By using pretrained models, researchers can take advantage of the general features learned by the model on a large dataset, such as ImageNet, and fine-tune it to perform well on a specific medical imaging task [76]. Transfer learning has been successfully applied in various medical imaging applications, including diagnosing diabetic retinopathy from retinal images, classifying skin cancer from dermoscopy images, and segmenting brain tumors from MRI scans [60,77,78,79].

2.2. Recurrent Neural Networks (RNNs)

2.2.1. Basic Concepts

RNNs are a class of DL models designed to handle sequential data [80]. Unlike feedforward neural networks, RNNs possess internal memory that enables them to maintain a hidden state across time steps, allowing them to learn patterns within sequences [81]. This property makes RNNs suitable for tasks that require processing time-dependent data, such as natural language processing, time-series prediction, and video analysis [82].

2.2.2. Architectures and Applications

While RNNs are less commonly used in medical imaging compared to CNNs, they have shown potential in specific applications that involve sequential data. Some well-known RNN architectures include the basic RNN (shown in Figure 3), long short-term memory (LSTM) [80], and gated recurrent unit (GRU) [81]. These architectures have been employed in various medical imaging tasks [83,84,85], such as image captioning, video analysis, and multimodal data fusion [86]. For instance, RNNs have been used in conjunction with CNNs for medical image captioning, where the goal is to generate a descriptive text for a given image [87]. In this context, a CNN is used to extract features from the image, while an RNN is employed to generate a sequence of words based on the extracted features [88]. This approach has been applied to generate radiology reports for chest X-rays and MRI scans [89]. Additionally, RNNs have been utilized for analyzing medical videos, such as endoscopy and laparoscopy videos [90]. In these applications, RNNs can be used to track and analyze temporal changes in the videos, enabling tasks such as surgical tool tracking, tissue segmentation, and surgical workflow analysis [55,91,92,93,94].

2.3. Generative Adversarial Networks (GANs)

2.3.1. Basic Concepts

Generative adversarial networks (GANs) has been regarded as a class of DL models designed for generating realistic samples from complex data distributions [95]. GANs consist of two neural networks, a generator and a discriminator, which are trained simultaneously in a minimax game [96]. The generator learns to produce samples that resemble the training data, while the discriminator learns to differentiate between real and generated samples. The training process continues until the generator produces samples that the discriminator cannot reliably distinguish from real data [97].

2.3.2. Architectures and Applications

Several GAN architectures and variants have been proposed for various tasks, including deep convolutional GAN (DCGAN) [98], Wasserstein GAN (WGAN) [99], and CycleGAN [100]. The basic mechanism of GANs has been shown in Figure 4. GANs have shown promising results in medical imaging applications [101,102,103], such as image synthesis, data augmentation, and image-to-image translation [104]. For example, GANs have been used to synthesize realistic medical images, which can be valuable for training other DL models, especially when annotated data is scarce [105]. In this context, GANs have been employed to generate synthetic CT, MRI, and ultrasound images, among others [106]. Another application of GANs in medical imaging is data augmentation, where GANs are used to create additional training samples to improve model performance and generalization [62]. By generating diverse and realistic variations of the available data, GANs can help mitigate issues related to limited datasets in medical imaging contexts [107]. Image-to-image translation is another application where GANs have shown potential in medical imaging. In this task, GANs are employed to transform images from one modality or representation to another, such as converting MRI images to CT images or enhancing image quality [108]. For instance, CycleGAN has been used for cross-modality synthesis between MRI and CT images, enabling the generation of synthetic CT images from MRI scans, which can be helpful in situations where CT scans are unavailable or contraindicated [63].

2.4. Limitations and Challenges

Despite the successes and potential of deep learning techniques in medical imaging, several common limitations and challenges need to be addressed. These challenges span across convolutional neural networks (CNNs), recurrent neural networks (RNNs), and generative adversarial networks (GANs). Shown in Table 1, one primary challenge is the lack of interpretability in deep learning models. CNNs, RNNs, and GANs often act as “black boxes,” making it difficult to understand the underlying decision-making processes. This lack of interpretability hinders their adoption in clinical practice, where explainability is crucial. Another challenge lies in the robustness and security of deep learning models. CNNs are susceptible to adversarial examples, which are carefully crafted inputs designed to deceive the model into making incorrect predictions. Adversarial attacks raise concerns about the reliability and trustworthiness of deep learning models in medical imaging applications. Furthermore, deep learning techniques, including CNNs, RNNs, and GANs, require large amounts of annotated data for training. Acquiring labeled medical imaging datasets can be time-consuming, expensive, and sometimes limited in size. Overcoming the challenge of data scarcity and finding efficient ways to leverage unlabeled data, such as unsupervised or semisupervised learning, is essential for the broader adoption of deep learning in medical imaging. Additionally, both RNNs and GANs face specific challenges. RNNs suffer from the vanishing and exploding gradient problem when training deep networks, making it difficult to learn long-term dependencies in sequences. The computational complexity of RNNs is also a concern, especially when dealing with long sequences or large-scale datasets. For GANs, the mode collapse problem is a significant challenge, as it can lead to limited variety and suboptimal results in tasks such as data augmentation and image synthesis. Training GANs can be challenging due to unstable dynamics and convergence issues. Ensuring the quality and reliability of generated images is crucial for their safe and effective use in medical imaging applications. Addressing these limitations and challenges will enhance the interpretability, robustness, scalability, and applicability of deep learning techniques in medical imaging.

3. Applications in Medical Imaging

To offer a comprehensive perspective on the role of DL in medical imaging, it is crucial to consider its multifaceted applications beyond just image reconstruction and registration. As highlighted in the title, the intention of this review is not merely to focus on these two aspects but to present a wider perspective on how DL is revolutionizing the field of medical imaging.
In the following sections, we delve into the specifics of how DL techniques have been employed in diverse tasks such as image segmentation and classification (Section 3.1 and Section 3.2), in addition to reconstruction and registration (Section 3.3 and Section 3.4). Image segmentation, for instance, involves partitioning a digital image into multiple segments to simplify the image and/or to extract relevant information. DL has significantly improved the performance of these tasks, making it a vital component of modern medical imaging. Similarly, image classification, which is the task of assigning an input image to one label from a fixed set of categories, is another area where DL has shown great potential. These varied applications underscore the breadth of DL’s impact on medical imaging, and it is this breadth that we seek to convey through this review.

3.1. Image Segmentation

3.1.1. Techniques and Approaches

Image segmentation is a critical task in medical imaging, which involves partitioning an image into multiple regions or segments, each representing a specific anatomical structure or region of interest (ROI) [45]. DL has shown exceptional performance in this domain, with CNNs being the most commonly used approach. The U-Net [6] is a popular CNN architecture specifically designed for biomedical image segmentation, which has been applied to various medical imaging modalities such as MRI, CT, and microscopy images [45]. Additionally, multiscale architectures [109], attention mechanisms [110], and 3D CNNs [111] have been proposed to improve segmentation accuracy and efficiency in complex medical imaging tasks. Figure 5 shows the application of DL approaches to liver segmentation.

3.1.2. Challenges and Future Directions

Despite the success of DL-based segmentation methods, several challenges remain. These include the need for large, annotated datasets, the limited interpretability of the models, and the robustness of the algorithms to variations in image quality, acquisition protocols, and patient populations [45,112]. Future research directions may focus on developing more efficient annotation techniques, incorporating domain knowledge into DL models, and improving the generalization capabilities of these models to unseen data or rare pathologies [113].

3.2. Image Classification

3.2.1. Techniques and Approaches

Image classification in medical imaging involves assigning a label to an input image, typically indicating the presence or absence of a specific condition or abnormality [114]. DL techniques, particularly CNNs, have demonstrated exceptional performance in image classification tasks [63]. Transfer learning, where pretrained models on large-scale natural image datasets (e.g., ImageNet) are fine-tuned on smaller medical imaging datasets, has been widely adopted to overcome the limitations of scarce labeled data in medical imaging [76]. Additionally, DL techniques such as DenseNets [68], ResNets [53], and multitask learning approaches [115] have been used to improve classification performance in various medical imaging applications, including the detection of cancerous lesions, identification of diseases, and assessment of treatment response. Figure 6 indicates the application of DL approach to mammography images.

3.2.2. Challenges and Future Directions

Key challenges in DL-based image classification include the limited availability of labeled data, class imbalance, and the need for model interpretability. Future research may focus on leveraging unsupervised or semisupervised learning techniques [116], data augmentation strategies [117], and advanced regularization techniques [118] to overcome these challenges. Moreover, developing methods to provide meaningful explanations for model predictions and incorporating domain knowledge into DL models may enhance their clinical utility [119].

3.3. Image Reconstruction

3.3.1. Techniques and Approaches

Image reconstruction is a fundamental step in many medical imaging modalities, such as CT, MRI, and PET, where raw data (e.g., projections, k-space data) are transformed into interpretable images [120]. DL has shown potential in improving image reconstruction quality and reducing reconstruction time [121]. CNNs have been used for image denoising, super-resolution, and artifact reduction in various imaging modalities [122,123]. Additionally, DL-based iterative reconstruction techniques [124] and the integration of DL models with conventional reconstruction algorithms [125] have been proposed to optimize image quality while reducing radiation dose or acquisition time. Figure 7 presents the application of GAN-based PET image reconstruction.

3.3.2. Challenges and Future Directions

Challenges in DL-based image reconstruction include the need for large-scale training data, the limited generalizability of the models across different imaging devices and acquisition protocols, and the potential for introducing new artifacts or biases into the reconstructed images [126]. Future research may focus on developing techniques to leverage limited training data, such as unsupervised or self-supervised learning methods [127], and designing more robust models that can generalize across different imaging conditions [128]. Furthermore, ensuring the safety and reliability of DL-based reconstruction methods by quantifying their uncertainties and validating their performance on large, diverse datasets will be crucial for their clinical adoption. Figure 8 is a typical example of DL-based medical image registration.

3.4. Image Registration

3.4.1. Techniques and Approaches

Image registration is the process of aligning two or more images, often acquired from different modalities or at different time points, to facilitate comparison and analysis [129]. DL has been increasingly applied to image registration tasks, with CNNs and spatial transformer networks (STNs) being the most commonly used architectures [130]. Supervised learning approaches, such as using ground-truth deformation fields or similarity metrics as labels, have been employed to train deep registration models [131]. Moreover, unsupervised learning techniques, which do not require ground-truth correspondences, have been proposed to overcome the challenges of obtaining labeled data for registration tasks [132].

3.4.2. Challenges and Future Directions

DL-based image registration faces challenges such as the need for large, diverse training datasets, the limited interpretability of the learned transformations, and the potential for overfitting or generating implausible deformations [70]. Future research may focus on developing more efficient and flexible DL architectures for registration, incorporating domain knowledge into the models, and designing robust evaluation metrics that can capture the clinical relevance of the registration results [133]. Additionally, leveraging multitask learning [134] and transfer learning approaches [135] may help improve the generalization and performance of deep registration models in various medical imaging applications.

4. Deep Learning for Specific Medical Imaging Modalities

Medical imaging modalities, such as magnetic resonance imaging (MRI), computed tomography (CT), positron emission tomography (PET), ultrasound, and optical coherence tomography (OCT), have unique characteristics and generate different types of images. Therefore, DL techniques need to be tailored to each modality to achieve optimal performance. In this section, we will discuss the current state-of-the-art DL techniques and applications for each modality, as well as the challenges and future directions.
Before diving into the application of DL in specific imaging modalities, it is important to clarify the focus of this section. The intention is to discuss how DL is applied in the analysis of images generated by these different modalities, such as MRI, CT, PET, ultrasound imaging, and OCT, rather than its application in the process of image acquisition. Specifically, the discussion will center around how DL has been utilized to extract meaningful insights from these images, for example, through tasks such as segmentation, classification, detection, and prediction. This includes the ability to identify and classify pathologies, measure anatomical structures, and even predict treatment outcomes.

4.1. Magnetic Resonance Imaging (MRI)

MRI is a noninvasive medical imaging modality that provides detailed structural and functional information. It has been widely used in diagnosis, treatment planning, and monitoring of various diseases. DL techniques have been applied to various tasks in MRI, including image segmentation, image registration, image synthesis, and disease classification.

4.1.1. DL Techniques and Applications to MRI

CNNs have been widely used in MRI analysis tasks. For instance, U-Net [6] has been used for MRI segmentation tasks, such as brain tumor segmentation [136] and prostate segmentation [137]. Similarly, residual networks (ResNets) [53] have been used for MRI reconstruction [119] and disease classification [110]. RNNs have also been used for MRI analysis, such as brain tumor segmentation [138]. GANs have also been used for MRI applications, such as image synthesis and image-to-image translation. For example, GANs have been used for the synthesis of brain MRI images [104] and for the generation of CT images from MRI images [124]. GANs have also been used for image denoising [139] and super-resolution [140] in MRI. Figure 9 shows the application of MRI brain image segmentation.

4.1.2. Challenges and Future Directions

Despite the promising results, there are still challenges in applying DL techniques to MRI analysis. One of the major challenges is the limited availability of large, annotated datasets. Moreover, the heterogeneity of MRI data, such as differences in image contrast, image resolution, and imaging protocols, makes it difficult to generalize DL models to new datasets. Therefore, developing transferable models that can handle these variations is an important future direction. Additionally, incorporating domain-specific knowledge and incorporating prior information into DL models can further improve their performance.

4.2. Computed Tomography (CT)

CT is a widely used medical imaging modality that provides detailed anatomical information. It is commonly used in the diagnosis and treatment planning of various diseases, such as cancer and cardiovascular diseases. DL techniques have been applied to various tasks in CT, including image segmentation, disease detection, and diagnosis.

4.2.1. DL Techniques and Applications to CT

CNNs have been widely used in CT analysis tasks. For example, Mask R-CNN [141] has been used for lung nodule detection in CT images [142]. CNNs have also been used for CT image registration [143] and image segmentation [144]. Moreover, DL techniques have been applied to CT angiography for vessel segmentation and centerline extraction [145]. In addition to CNNs, GANs have also been used in CT applications, such as image denoising [146] and image synthesis [147]. For example, GANs have been used for synthesizing low-dose CT images from high-dose CT images [148]. Figure 10 indicates CT image classification using DL techniques.

4.2.2. Challenges and Future Directions

One of the challenges in applying DL techniques to CT analysis is the limited availability of annotated datasets. Moreover, CT images contain high levels of noise, which can affect the performance of DL models. Therefore, developing DL models that are robust to noise is an important future direction. Moreover, developing transferable models that can handle variations in imaging protocols and patient populations is also an important future direction.

4.3. Positron Emission Tomography (PET)

PET is a medical imaging modality that is used for functional imaging. It is commonly used in cancer diagnosis and treatment planning. DL techniques have been applied to various tasks in PET, including image segmentation, image reconstruction, and disease classification. Figure 11 is a typical example of PET image segmentation using a DL-based method.

4.3.1. DL Techniques and Applications to PET

CNNs have been widely used in PET image analysis tasks. For example, U-Net [6] has been used for PET image segmentation [149]. Moreover, GANs have been used for PET image reconstruction [150] and image denoising [151]. Additionally, DL techniques have been applied to PET image registration [152] and disease classification [140].

4.3.2. Challenges and Future Directions

One of the challenges in applying DL techniques to PET is the limited availability of annotated datasets, particularly for rare diseases. Moreover, PET images suffer from low spatial resolution and high noise levels, which can affect the performance of DL models. Therefore, developing robust DL models that can handle these challenges is an important future direction. Additionally, developing transferable models that can handle variations in imaging protocols and patient populations is also an important future direction.

4.4. Ultrasound Imaging

Ultrasound is a medical imaging modality that uses high-frequency sound waves to produce images of the internal organs and tissues. It is commonly used in obstetrics, cardiology, and urology. DL techniques have been applied to various tasks in ultrasound imaging, including image segmentation, disease classification, and image denoising. Figure 12 presents one example of fetal head detection in ultrasound images using convolutional neural networks.

4.4.1. DL Techniques and Applications to Ultrasound

CNNs have been widely used in ultrasound image analysis tasks. For example, U-Net [6] has been used for segmentation of the fetal brain in ultrasound images [153]. Moreover, RNNs have been used for tracking the fetal brain in ultrasound videos [154]. Additionally, DL techniques have been applied to ultrasound elastography for tissue characterization [155].

4.4.2. Challenges and Future Directions

One of the challenges in applying DL techniques to ultrasound imaging is the limited availability of annotated datasets, particularly for rare diseases. Moreover, ultrasound images are prone to artifacts and noise, which can affect the performance of DL models. Therefore, developing robust DL models that can handle these challenges is an important future direction. Additionally, developing transferable models that can handle variations in imaging protocols and patient populations is also an important future direction.

4.5. Optical Coherence Tomography (OCT)

OCT is a medical imaging modality that uses light waves to produce images of biological tissues. It is commonly used in ophthalmology for imaging the retina and the optic nerve. DL techniques have been applied to various tasks in OCT imaging, including image segmentation, disease classification, and image registration. Figure 13 is the workflow of OCT image angiography using a DL-based approach.

4.5.1. DL Techniques and Applications to OCT

CNNs have been widely used in OCT image analysis tasks. For example, a fully convolutional network (FCN) has been used for segmentation of retinal layers in OCT images [156]. Moreover, DL techniques have been applied to OCT angiography for vessel segmentation and centerline extraction [6]. Additionally, RNNs have been used for tracking the movement of retinal layers in OCT videos [44].

4.5.2. Challenges and Future Directions

One of the challenges in applying DL techniques to OCT imaging is the limited availability of annotated datasets, particularly for rare diseases. Moreover, OCT images suffer from speckle noise and low signal-to-noise ratio, which can affect the performance of DL models. Therefore, developing robust DL models that can handle these challenges is an important future direction. Additionally, developing transferable models that can handle variations in imaging protocols and patient populations is also an important future direction.

5. Evaluation Methods and Available Datasets

DL techniques for medical imaging have shown impressive performance in various tasks, including image segmentation, classification, reconstruction, and registration. To evaluate the performance of these methods, appropriate metrics and benchmarks are needed.

5.1. Metrics for Performance Evaluation

Various metrics have been proposed to evaluate the performance of DL methods for medical imaging. For image segmentation tasks, commonly used metrics include Dice coefficient, Jaccard index, and surface distance measures [157]. For image classification tasks, metrics such as accuracy, precision, recall, and F1 score are often used [158]. For image reconstruction tasks, peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) are commonly used metrics [159]. In addition, some studies have proposed novel metrics specific to certain applications, such as registration accuracy and tumor size measurement in cancer imaging [160]. It is important to note that no single metric can fully capture the performance of a DL method, and a combination of metrics should be used for comprehensive evaluation. Moreover, the choice of metrics should depend on the specific application and clinical relevance.
In the realm of medical imaging, various deep learning (DL) methods have been applied and compared in terms of their performance. For instance, a notable comparative study by Zhang et al. [161] explored the effectiveness of convolutional neural networks (CNNs) and recurrent neural networks (RNNs) in detecting tumors from lung CT scans. The study revealed that both models yielded commendable results, but CNNs outshined RNNs with an accuracy rate of 92% compared to 89%. Furthermore, the CNN model demonstrated superior sensitivity and specificity, underscoring the potential advantages of CNNs in tasks involving medical imaging. Another insightful comparison was presented by Patel et al. [162], which contrasted the performance of deep belief networks (DBNs) and CNNs for the detection of breast cancer using mammograms. Although both models achieved impressive accuracy rates, the DBN demonstrated a superior area under the receiver operating characteristic (ROC) curve (AUC), scoring 0.96 compared to the CNN’s 0.92. This finding suggests that DBNs might offer an edge over CNNs in distinguishing between malignant and benign cases in mammography.

5.2. Publicly Available Datasets and Competitions

Publicly available datasets and competitions play a critical role in advancing DL research for medical imaging. These resources provide standardized data and evaluation protocols for comparing different methods and fostering collaboration among researchers. There are various publicly available datasets for different medical imaging modalities, such as the Cancer Imaging Archive (TCIA) for CT and MRI, the Alzheimer’s Disease Neuroimaging Initiative (ADNI) for MRI [163], and the Retinal OCT (ORIGA) dataset for OCT [164]. In addition, several competitions have been organized to benchmark the performance of DL methods for medical imaging, such as the International Symposium on Biomedical Imaging (ISBI) challenge [165] and the Medical Segmentation Decathlon [166]. However, the availability and quality of publicly available datasets and competitions can vary across different medical imaging modalities and tasks. Moreover, some datasets may have limited diversity in terms of patient populations and imaging protocols, which can affect the generalizability of the results. To address these issues, it is important to establish standards and guidelines for dataset curation and evaluation protocols. Collaborative efforts among researchers, clinicians, and industry partners are needed to ensure the availability and quality of publicly available datasets and competitions for DL research in medical imaging.

6. Ethical Considerations for Using DL Methods

In recent years, the rapid development and widespread use of DL techniques in medical imaging have raised a number of ethical considerations, ranging from data privacy and security to bias and fairness, explainability and interpretability, and integration with clinical workflows. In this section, we discuss some of these issues and their potential impact on the future of DL in medical imaging.

6.1. Data Privacy and Security

One of the main ethical concerns associated with DL in medical imaging is the need to protect patient data privacy and ensure data security. Medical images contain sensitive information about patients, and their unauthorized use or disclosure could have serious consequences for their privacy and well-being. Therefore, it is essential to implement appropriate measures to protect the confidentiality, integrity, and availability of medical images and associated data. Several studies have proposed various methods for enhancing data privacy and security in medical imaging, including encryption, anonymization, and secure data sharing protocols [167]. These methods can help to protect patient data privacy and reduce the risk of data breaches or cyberattacks.

6.2. Bias and Fairness

Another important ethical consideration in the use of DL in medical imaging is the risk of bias and unfairness. DL models are trained on large datasets, and if these datasets are biased or unrepresentative, the resulting models can perpetuate or amplify these biases, leading to unfair or inaccurate predictions [168]. Several studies have highlighted the issue of bias in medical imaging datasets, such as disparities in the representation of certain demographic groups [169]. To address these issues, researchers have proposed various approaches, such as data augmentation, data balancing, and fairness-aware training [77]. These methods can help to mitigate bias and improve the fairness of DL models.

6.3. Explainability and Interpretability

The black-box nature of DL models is another ethical concern in medical imaging, as it can make it difficult to understand how they arrive at their predictions, and to identify potential errors or biases [170]. This lack of transparency and interpretability can limit the usefulness of DL in clinical settings, where explainability and interpretability are critical for building trust and confidence among healthcare providers and patients. To address these issues, researchers have proposed various methods for enhancing the explainability and interpretability of DL models, such as attention mechanisms, saliency maps, and counterfactual explanations [44]. These methods can help to improve the transparency and interpretability of DL models and facilitate their integration into clinical workflows.

6.4. Integration with Clinical Workflows

The integration of DL into clinical workflows is another important consideration in the use of DL in medical imaging. To be clinically useful, DL models must be integrated into clinical workflows in a way that is efficient, reliable, and effective [171]. This requires careful consideration of various factors, such as the availability and accessibility of data, the quality and relevance of predictions, and the impact on clinical decision-making. Several studies have proposed various methods for integrating DL into clinical workflows, such as decision support systems, clinical decision rules, and workflow optimization [172]. These methods can help to streamline the use of DL in clinical settings and improve the efficiency and effectiveness of clinical decision-making.

6.5. Future Research Directions

Looking forward, there are several key areas for future research in the use of DL in medical imaging. These include the following: (1) Developing more robust and accurate DL models that can handle variations in data quality and heterogeneity. (2) Enhancing the interpretability and explainability of DL models to facilitate their integration into clinical workflows. (3) Addressing ethical considerations, such as data privacy and security, bias and fairness, and regulatory compliance. (4) Investigating the potential of using DL in combination with other modalities, such as genomics, proteomics, and metabolomics, to improve the accuracy and specificity of medical imaging diagnoses. (5) Exploring the use of DL in personalized medicine, where models can be trained on patient-specific data to provide tailored treatment recommendations. (6) Developing methods for ensuring the robustness and generalizability of DL models across different populations and clinical settings. (7) Investigating the potential of using DL to automate the entire medical imaging pipeline, from acquisition to analysis to interpretation.
In conclusion, DL techniques have shown great promise in the field of medical imaging, with a wide range of applications and potential benefits for patient care. However, their use also raises important ethical considerations, such as data privacy and security, bias and fairness, and explainability and interpretability. Addressing these issues will be critical to realizing the full potential of DL in medical imaging and ensuring that its benefits are equitably distributed. Future research should focus on developing more robust and accurate models, enhancing their interpretability and explainability, and exploring new applications and use cases for DL in medical imaging. Moreover, it is important to collaborate with healthcare providers, patients, and other stakeholders to ensure that the development and use of DL models in medical imaging align with their needs and priorities. This includes involving patients in the design and evaluation of DL models and ensuring that the benefits of these models are accessible to all, regardless of socioeconomic status, race, or ethnicity. In addition, regulatory frameworks must be established to ensure that DL models meet ethical and quality standards and that their use is transparent and accountable. This includes developing guidelines for data privacy and security, bias and fairness, and explainability and interpretability, as well as establishing standards for model validation and performance evaluation. Overall, DL has the potential to revolutionize the field of medical imaging and transform the way we diagnose and treat diseases. However, its success will depend on addressing the ethical and technical challenges that come with its use and on developing a collaborative and patient-centered approach to its development and implementation. With continued research and innovation, DL is poised to make a significant contribution to the advancement of healthcare and improve the lives of patients around the world.

7. Conclusions

In this review article, we provided a comprehensive analysis of DL techniques and their applications in the field of medical imaging. We discussed the impact of DL on disease diagnosis and treatment and how it has transformed the medical imaging landscape. Furthermore, we reviewed the most recent DL techniques, such as CNNs, RNNs, and GANs, and their applications in medical imaging.
We explored the application of DL in various medical imaging modalities, including MRI, CT, PET, ultrasound imaging, and OCT. We also discussed the evaluation metrics and benchmarks used to assess the performance of DL algorithms in medical imaging, as well as the ethical considerations and future perspectives of the field.
Moving forward, the integration of DL with medical imaging is expected to continue revolutionizing the diagnosis, treatment, and management of diseases. The development of more advanced algorithms, coupled with the ever-increasing availability of medical imaging data, will undoubtedly contribute to significant advancements in healthcare. However, the medical community must also address the various challenges and ethical considerations that arise in the application of DL, such as data privacy, security, bias, and interpretability, to ensure that the technology is responsibly harnessed for the betterment of patient care.
Overall, DL in medical imaging holds great promise for improving healthcare outcomes and advancing the field of medicine. As the technology continues to evolve, it is essential for researchers, clinicians, and other stakeholders to work collaboratively to overcome challenges, address ethical concerns, and fully realize the potential of DL in medical imaging.

Author Contributions

Conceptualization, H.Z. and Y.Q.; methodology, H.Z. and Y.Q.; software, H.Z. and Y.Q.; validation, H.Z. and Y.Q.; formal analysis, H.Z. and Y.Q.; investigation, H.Z. and Y.Q.; resources, H.Z. and Y.Q.; data curation, H.Z. and Y.Q.; writing—original draft preparation, H.Z. and Y.Q.; writing—review and editing, H.Z. and Y.Q.; visualization, H.Z. and Y.Q.; supervision, H.Z.; project administration, H.Z. and Y.Q.; funding acquisition, H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the National Natural Science Foundation of China under Grant 62127812, 61971335.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are available upon request by email to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ayache, N. Medical Imaging in the Age of Artificial Intelligence. Healthc. Artif. Intell. 2020, 89–91. [Google Scholar]
  2. Wang, W.; Liang, D.; Chen, Q.; Iwamoto, Y.; Han, X.H.; Zhang, Q.; Hu, H.; Lin, L.; Chen, Y.W. Medical image classification using deep learning. Deep. Learn. Healthc. Paradig. Appl. 2020, 33–51. [Google Scholar]
  3. Fourcade, A.; Khonsari, R.H. Deep learning in medical image analysis: A third eye for doctors. J. Stomatol. Oral Maxillofac. Surg. 2019, 120, 279–288. [Google Scholar] [CrossRef] [PubMed]
  4. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  5. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  6. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. U-Net: Convolutional networks for biomedical image segmentation. Adv. Neural Inf. Process. Syst. 2002, 25, 1097–1105. [Google Scholar]
  7. Yao, H.M.; Sha, W.E.I.; Jiang, L. Two-Step Enhanced Deep Learning Approach for Electromagnetic Inverse Scattering Problems. IEEE Antennas Wirel. Propag. Lett. 2019, 18, 2254–2258. [Google Scholar] [CrossRef]
  8. Yao, H.M.; Jiang, L.; Wei, E.I. Enhanced Deep Learning Approach Based on the Deep Convolutional Encod-er-Decoder Architecture for Electromagnetic Inverse Scattering Problems. IEEE Antennas Wirel. Propag. Lett. 2020, 19, 1211–1215. [Google Scholar] [CrossRef]
  9. Guo, R.; Li, C.; Chen, X.; Yang, J.; Zhang, B.; Cheng, Y. Joint inversion of audio-magnetotelluric and seismic travel time data with deep learning constraint. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7982–7995. [Google Scholar] [CrossRef]
  10. Yao, H.M.; Guo, R.; Li, M.; Jiang, L.; Ng, M.K.P. Enhanced Supervised Descent Learning Technique for Electromagnetic Inverse Scattering Problems by the Deep Convolutional Neural Networks. IEEE Trans. Antennas Propag. 2022, 70, 6195–6206. [Google Scholar] [CrossRef]
  11. Yao, H.M.; Jiang, L. Enhanced PML Based on the Long Short Term Memory Network for the FDTD Method. IEEE Access 2020, 8, 21028–21035. [Google Scholar] [CrossRef]
  12. Yao, H.M.; Jiang, L.; Ng, M. Implementing the Fast Full-Wave Electromagnetic Forward Solver Using the Deep Convolutional Encoder-Decoder Architecture. IEEE Trans. Antennas Propag. 2022, 71, 1152–1157. [Google Scholar] [CrossRef]
  13. Zhang, H.H.; Yao, H.M.; Jiang, L.; Ng, M. Solving Electromagnetic Inverse Scattering Problems in Inhomogeneous Media by Deep Convolutional Encoder–Decoder Structure. IEEE Trans. Antennas Propag. 2023, 71, 2867–2872. [Google Scholar] [CrossRef]
  14. Zhang, H.H.; Yao, H.M.; Jiang, L.; Ng, M. Enhanced Two-Step Deep-Learning Approach for Electromagnetic-Inverse-Scattering Problems: Frequency Extrapolation and Scatterer Reconstruction. IEEE Trans. Antennas Propag. 2022, 71, 1662–1672. [Google Scholar] [CrossRef]
  15. Zhang, H.H.; Li, J.; Yao, H.M. Fast Full Wave Electromagnetic Forward Solver Based on Deep Conditional Convolutional Autoencoders. IEEE Antennas Wirel. Propag. Lett. 2022, 22, 779–783. [Google Scholar] [CrossRef]
  16. Zhang, H.H.; Li, J.; Yao, H.M. Deep Long Short-Term Memory Networks-Based Solving Method for the FDTD Method: 2-D Case. IEEE Microw. Wirel. Technol. Lett. 2023, 33, 499–502. [Google Scholar] [CrossRef]
  17. Yao, H.M.; Jiang, L. Machine-Learning-Based PML for the FDTD Method. IEEE Antennas Wirel. Propag. Lett. 2018, 18, 192–196. [Google Scholar] [CrossRef]
  18. Yao, H.; Zhang, L.; Yang, H.; Li, M.; Zhang, B. Snow Parameters Inversion from Passive Microwave Remote Sensing Measurements by Deep Convolutional Neural Networks. Sensors 2022, 22, 4769. [Google Scholar] [CrossRef] [PubMed]
  19. Yao, H.M.; Sha, W.E.I.; Jiang, L.J. Applying Convolutional Neural Networks for The Source Reconstruction. Prog. Electromagn. Res. M 2018, 76, 91–99. [Google Scholar] [CrossRef]
  20. Yao, H.M.; Li, M.; Jiang, L. Applying Deep Learning Approach to the Far-Field Subwavelength Imaging Based on Near-Field Resonant Metalens at Microwave Frequencies. IEEE Access 2019, 7, 63801–63808. [Google Scholar] [CrossRef]
  21. Zhang, H.H.; Jiang, L.; Yao, H.M. Embedding the behavior macromodel into TDIE for transient field-circuit simulations. IEEE Trans. Antennas Propag. 2016, 64, 3233–3238. [Google Scholar] [CrossRef]
  22. Zhang, H.H.; Jiang, L.J.; Yao, H.M.; Zhang, Y. Transient Heterogeneous Electromagnetic Simulation with DGTD and Behavioral Macromodel. IEEE Trans. Electromagn. Compat. 2017, 59, 1152–1160. [Google Scholar] [CrossRef]
  23. Xiao, B.; Yao, H.; Li, M.; Hong, J.S.; Yeung, K.L. Flexible Wideband Microstrip-Slotline-Microstrip Power Divider and Its Application to Antenna Array. IEEE Access 2019, 7, 143973–143979. [Google Scholar] [CrossRef]
  24. Li, M.; Wang, R.; Yao, H.; Wang, B. A Low-Profile Wideband CP End-Fire Magnetoelectric Antenna Using Dual-Mode Resonances. IEEE Trans. Antennas Propag. 2019, 67, 4445–4452. [Google Scholar] [CrossRef]
  25. Yao, H.M.; Jiang, L.; Zhang, H.H.; Wei, E.I. Machine learning methodology review for computational electromagnetics. In Proceedings of the 2019 International Applied Computational Electromagnetics Society Symposium-China (ACES), Washington, DC, USA, 10–13 October 2019; Volume 1. [Google Scholar]
  26. Guo, R.; Li, M.; Yang, F.; Yao, H.; Jiang, L.; Ng, M.; Abubakar, A. Joint 2D inversion of AMT and seismic traveltime data with deep learning constraint. In Proceedings of the SEG International Exposition and Annual Meeting, Virtual, 11–16 October 2020. [Google Scholar] [CrossRef]
  27. Yao, H.M.; Jiang, L.J.; Qin, Y.W. Machine learning based method of moments (ML-MoM). In Proceedings of the 2017 IEEE International Symposium on Antennas and Propagation & USNC/URSI National Radio Science Meeting, San Diego, CA, USA, 9–14 July 2017. [Google Scholar]
  28. Yao, H.M.; Qin, Y.W.; Jiang, L.J. Machine learning based MoM (ML-MoM) for parasitic capacitance extractions. In Proceedings of the 2016 IEEE Electrical Design of Advanced Packaging and Systems (EDAPS), Honolulu, HI, USA, 14–16 December 2016. [Google Scholar]
  29. Yao, H.M.; Jiang, L.J. Machine learning based neural network solving methods for the FDTD method. In Proceedings of the 2018 IEEE International Symposium on Antennas and Propagation & USNC/URSI National Radio Science Meeting, Boston, MA, USA, 8–13 July 2018. [Google Scholar]
  30. Jiang, L.; Yao, H.; Zhang, H.; Qin, Y. Machine Learning Based Computational Electromagnetic Analysis for Electromagnetic Compatibility. In Proceedings of the 2018 IEEE International Conference on Computational Electromagnetics (ICCEM), Chengdu, China, 26–28 March 2018. [Google Scholar]
  31. Yao, H.M.; Jiang, L.J.; Wei, E.I. Source Reconstruction Method based on Machine Learning Algorithms. In Proceedings of the 2019 Joint International Symposium on Electromagnetic Compatibility, Sapporo and Asia-Pacific International Symposium on Electromagnetic Compatibility (EMC Sapporo/APEMC), Sapporo, Japan, 3–7 June 2019. [Google Scholar]
  32. Zhang, H.H.; Yao, H.M.; Jiang, L.J. Novel time domain integral equation method hybridized with the macromodels of circuits. In Proceedings of the 2015 IEEE 24th Electrical Performance of Electronic Packaging and Systems (EPEPS), San Jose, CA, USA, 25–28 October 2015. [Google Scholar]
  33. Zhang, H.H.; Jiang, L.J.; Yao, H.M.; Zhang, Y. Coupling DGTD and behavioral macromodel for transient heterogeneous electromagnetic simulations. In Proceedings of the 2016 IEEE International Symposium on Electromagnetic Compatibility (EMC), Ottawa, ON, Canada, 25–29 July 2016. [Google Scholar]
  34. Zhang, H.H.; Jiang, L.J.; Yao, H.M.; Zhao, X.W.; Zhang, Y. Hybrid field-circuit simulation by coupling DGTD with behavioral macromodel. In Proceedings of the 2016 Progress in Electromagnetic Research Symposium (PIERS), Shanghai, China, 8–11 August 2016. [Google Scholar]
  35. Yao, H.; Hsieh, Y.-P.; Kong, J.; Hofmann, M. Modelling electrical conduction in nanostructure assemblies through complex networks. Nat. Mater. 2020, 19, 745–751. [Google Scholar] [CrossRef] [PubMed]
  36. Yao, H.; Hempel, M.; Hsieh, Y.-P.; Kong, J.; Hofmann, M. Characterizing percolative materials by straining. Nanoscale 2018, 11, 1074–1079. [Google Scholar] [CrossRef] [PubMed]
  37. Guo, S.; Fu, J.; Zhang, P.; Zhu, C.; Yao, H.; Xu, M.; An, B.; Wang, X.; Tang, B.; Deng, Y.; et al. Direct growth of single-metal-atom chains. Nat. Synth. 2022, 1, 245–253. [Google Scholar] [CrossRef]
  38. Liu, H.; Yao, H.; Feng, L. A nanometer-resolution displacement measurement system based on laser feedback interferometry. In Proceedings of the 8th Annual IEEE International Conference on Nano/Micro Engineered and Molecular Systems, Xiamen, China, 7–10 April 2013. [Google Scholar]
  39. Liu, H.L.; Yao, H.M.; Meng, Z.K.; Feng, L.S. Simulation and Error Analysis of a Laser Feedback Interference System Based on Phase-freezing Technology. Lasers Eng. 2014, 29, 259–270. [Google Scholar]
  40. Chen, D.-R.; Hofmann, M.; Yao, H.-M.; Chiu, S.-K.; Chen, S.-H.; Luo, Y.-R.; Hsu, C.-C.; Hsieh, Y.-P. Lateral Two-Dimensional Material Heterojunction Photodetectors with Ultrahigh Speed and Detectivity. ACS Appl. Mater. Interfaces 2019, 11, 6384–6388. [Google Scholar] [CrossRef]
  41. Chen, Z.; Ming, T.; Goulamaly, M.M.; Yao, H.; Nezich, D.; Hempel, M.; Hofmann, M.; Kong, J. Enhancing the Sensitivity of Percolative Graphene Films for Flexible and Transparent Pressure Sensor Arrays. Adv. Funct. Mater. 2016, 26, 5061–5067. [Google Scholar] [CrossRef]
  42. Yao, H.M.; Li, M.; Jiang, L.; Ng, M. Antenna Array Diagnosis by Using Deep Learning Approach. IEEE Trans. Antennas Propag. 2023. early access. [Google Scholar]
  43. Yao, H.M.; Jiang, L.; Ng, M. Enhanced Deep Learning Approach Based on the Conditional Generative Adversarial Network for Electromagnetic Inverse Scattering Problems. IEEE Trans. Antennas Propag. 2023. early access. [Google Scholar]
  44. Shen, D.; Wu, G.; Suk, H.-I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [PubMed]
  45. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed]
  46. Anaya-Isaza, A.; Mera-Jiménez, L.; Zequera-Diaz, M. An overview of deep learning in medical imaging. Inform. Med. Unlocked 2021, 26, 100723. [Google Scholar] [CrossRef]
  47. Ghaderzadeh, M.; Asadi, F. Deep Learning in the Detection and Diagnosis of COVID-19 Using Radiology Modalities: A Sys-tematic Review. J. Healthc. Eng. 2021, 2021, 6677314. [Google Scholar] [CrossRef] [PubMed]
  48. Ghaderzadeh, M.; Asadi, F.; Jafari, R.; Bashash, D.; Abolghasemi, H.; Aria, M. Deep Convolutional Neural Network–Based Computer-Aided Detection System for COVID-19 Using Multiple Lung Scans: Design and Implementation Study. J. Med. Internet Res. 2021, 23, e27468. [Google Scholar] [CrossRef] [PubMed]
  49. Ghaderzadeh, M.; Aria, M.; Hosseini, A.; Asadi, F.; Bashash, D.; Abolghasemi, H. A fast and efficient CNN model for B-ALL diagnosis and its subtypes classification using peripheral blood smear images. Int. J. Intell. Syst. 2021, 37, 5113–5133. [Google Scholar] [CrossRef]
  50. Ghaderzadeh, M.; Aria, M.; Asadi, F. X-Ray Equipped with Artificial Intelligence: Changing the COVID-19 Diagnostic Para-digm During the Pandemic. BioMed Res. Int. 2021, 2021, 9942873. [Google Scholar] [CrossRef]
  51. Ghaderzadeh, M.; Aria, M. Management of COVID-19 Detection Using Artificial Intelligence in 2020 Pandemic. In Proceedings of the 5th International Conference on Medical and Health Informatics, Kyoto, Japan, 14–16 May 2021. [Google Scholar]
  52. Gheisari, M.; Ebrahimzadeh, F.; Rahimi, M.; Moazzamigodarzi, M.; Liu, Y.; Pramanik, P.K.D.; Heravi, M.A.; Mehbodniya, A.; Ghaderzadeh, M.; Feylizadeh, M.R.; et al. Deep learning: Applications, architectures, models, tools, and frameworks: A comprehensive survey. CAAI Trans. Intell. Technol. 2023. early view. [Google Scholar] [CrossRef]
  53. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  54. Shah, S.A.A.; Tahir, S.A.; Aksam Iftikhar, M. A comprehensive survey on deep learning-based approaches for medical image analysis. Comput. Electr. Eng. 2021, 90, 106954. [Google Scholar]
  55. Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  56. Anholt, W.J.H.V.; Dankelman, J.W.; Wauben, L.S.G.L. An overview of medical imaging modalities: The role of imaging physics in medical education. Eur. J. Phys. Educ. 2020, 11, 12–28. [Google Scholar]
  57. A. C. Society. Imaging (Radiology) Tests, American Cancer Society. 2018. Available online: https://www.cancer.org/treatment/understanding-your-diagnosis/tests/imaging-radiology-tests-for-cancer.html (accessed on 23 May 2023).
  58. Simpson, A.L.; Antonelli, M.; Bakas, S.; Bilello, M.; Farahani, K.; Van Ginneken, B.; Kopp-Schneider, A.; Landman, B.A.; Litjens, G.; Menze, B.; et al. A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv 2019, arXiv:1902.09063. [Google Scholar]
  59. Sweeney, G.J. Big data, big problems: Emerging issues in the ethics of data science and journalism. J. Mass Media Ethics 2014, 29, 38–51. [Google Scholar]
  60. Lipton, Z.C. The Mythos of Model Interpretability. Queue 2018, 16, 31–57. [Google Scholar] [CrossRef]
  61. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  62. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  63. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  64. Luo, X.; Hu, M.; Song, T.; Wang, G.; Zhang, S. Semi-supervised medical image segmentation via cross teaching between CNN and transformer. In Proceedings of the International Conference on Medical Imaging with Deep Learning, PMLR, Durham, NC, USA, 5–6 August 2022. [Google Scholar]
  65. Tiwari, P.; Pant, B.; Elarabawy, M.M.; Abd-Elnaby, M.; Mohd, N.; Dhiman, G.; Sharma, S. CNN Based Multiclass Brain Tumor Detection Using Medical Imaging. Comput. Intell. Neurosci. 2022, 2022, 1830010. [Google Scholar] [CrossRef] [PubMed]
  66. Srikantamurthy, M.M.; Rallabandi, V.P.; Dudekula, D.B.; Natarajan, S.; Park, J. Classification of benign and malignant subtypes of breast cancer histopathology imaging using hybrid CNN-LSTM based transfer learning. BMC Med. Imaging 2023, 23, 19. [Google Scholar] [CrossRef]
  67. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  68. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  69. Greenspan, H.; van Ginneken, B.; Summers, R.M. Guest Editorial Deep Learning in Medical Imaging: Overview and Future Promise of an Exciting New Technique. IEEE Trans. Med. Imaging 2016, 35, 1153–1159. [Google Scholar] [CrossRef]
  70. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  71. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. DataEng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  72. Awad, M.M. Evaluation of COVID-19 Reported Statistical Data Using Cooperative Convolutional Neural Network Model (CCNN). COVID 2022, 2, 674–690. [Google Scholar] [CrossRef]
  73. Li, Z.; Zhang, H.; Li, Z.; Ren, Z. Residual-Attention UNet++: A Nested Residual-Attention U-Net for Medical Image Segmentation. Appl. Sci. 2022, 12, 7149. [Google Scholar] [CrossRef]
  74. Safarov, S.; Whangbo, T.K. A-DenseUNet: Adaptive Densely Connected UNet for Polyp Segmentation in Colonoscopy Images with Atrous Convolution. Sensors 2021, 21, 1441. [Google Scholar] [CrossRef] [PubMed]
  75. Khan, S.; Rahmani, H.; Shah, S.A.A.; Bennamoun, M. A Guide to Convolutional Neural Networks for Computer Vision; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar] [CrossRef]
  76. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef]
  77. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  78. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  79. Zikic, D.; Glocker, B.; Konukoglu, E.; Criminisi, A.; Demiralp, C.; Shotton, J.; Thomas, O.M.; Das, T.; Jena, R.; Price, S.J. Decision Forests for Tissue-Specific Segmentation of High-Grade Gliomas in Multi-channel MR. In Proceedings of the MICCAI 2012, Nice, France, 1–5 October 2012; Volume 15, pp. 369–376. [Google Scholar] [CrossRef]
  80. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  81. Cho, K.; Van Merrienboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar] [CrossRef]
  82. Donahue, J.; Hendricks, L.A.; Guadarrama, S.; Rohrbach, M.; Venugopalan, S.; Saenko, K.; Darrell, T. Long-term Recurrent Convolutional Networks for Visual Recognition and Description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar] [CrossRef]
  83. Sridhar, C.; Pareek, P.K.; Kalidoss, R.; Jamal, S.S.; Shukla, P.K.; Nuagah, S.J. Optimal Medical Image Size Reduction Model Creation Using Recurrent Neural Network and GenPSOWVQ. J. Health Eng. 2022, 2022, 1–8. [Google Scholar] [CrossRef] [PubMed]
  84. Chen, E.Z.; Wang, P.; Chen, X.; Chen, T.; Sun, S. Pyramid Convolutional RNN for MRI Image Reconstruction. IEEE Trans. Med. Imaging 2022, 41, 2033–2047. [Google Scholar] [CrossRef] [PubMed]
  85. Suganyadevi, S.; Seethalakshmi, V.; Balasamy, K. A review on deep learning in medical image analysis. Int. J. Multimed. Inf. Retr. 2022, 11, 19–38. [Google Scholar] [CrossRef] [PubMed]
  86. Setio, A.A.; Ciompi, F.; Litjens, G.; Gerke, P.; Jacobs, C.; van Riel, S.J.; Wille, M.M.; Naqibullah, M.; Sanchez, C.I.; van Ginneken, B. Pulmonary Nodule Detection in CT Images: False Positive Reduction Using Multi-View Convolutional Networks. IEEE Trans. Med. Imaging 2016, 35, 1160–1169. [Google Scholar] [CrossRef] [PubMed]
  87. Yang, K.; Mohammed, E.A.; Far, B.H. Detection of Alzheimer’s Disease Using Graph-Regularized Convolutional Neural Network Based on Structural Similarity Learning of Brain Magnetic Resonance Images. In Proceedings of the 2021 IEEE 22nd International Conference on Information Reuse and Integration for Data Science (IRI), Las Vegas, NV, USA, 10–12 August 2021; pp. 326–333. [Google Scholar]
  88. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar]
  89. Li, Z.; Wang, C.; Han, M.; Xue, Y.; Wei, W.; Li, L.J.; Fei-Fei, L. Thoracic disease identification and localization with limited supervision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8290–8299. [Google Scholar]
  90. Padoy, N. Towards automatic recognition of surgical activities. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Nice, France, 1–5 October 2012; pp. 267–274. [Google Scholar]
  91. Mutter, V.; Gangi, A.; Rekik, M.A. A survey of deep learning techniques for medical image segmentation. In Deep Learning and Convolutional Neural Networks for Medical Imaging and Clinical Informatics; Springer: Berlin/Heidelberg, Germany, 2019; pp. 21–45. [Google Scholar]
  92. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Advances in Neural Information Processing Systems; Springer: Berlin/Heidelberg, Germany, 2014; pp. 2672–2680. [Google Scholar]
  93. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  94. Schlegl, T.; Seeböck, P.; Waldstein, S.M.; Schmidt-Erfurth, U.; Langs, G. Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide Marker Discovery. In Proceedings of the International Conference on Information Processing in Medical Imaging, Boone, NC, USA, 25–30 June 2017; pp. 146–157. [Google Scholar]
  95. Frid-Adar, M.; Diamant, I.; Klang, E.; Amitai, M.; Goldberger, J.; Greenspan, H. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion clas-sification. Neurocomputing 2018, 321, 321–331. [Google Scholar] [CrossRef]
  96. Han, Y. MR-based synthetic CT generation using a deep convolutional neural network method. Med. Phys. 2017, 44, 1408–1419. [Google Scholar] [CrossRef]
  97. Yi, X.; Walia, E.; Babyn, P. Generative adversarial network in medical imaging: A review. Med. Image Anal. 2019, 58, 101552. [Google Scholar] [CrossRef]
  98. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.P.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  99. Chen, Y.; Xie, Y.; Zhou, Z.; Shi, F.; Christodoulou, A.G.; Li, D. Brain MRI super resolution using 3D deep densely connected neural networks. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 739–742. [Google Scholar]
  100. Choi, Y.; Choi, M.; Kim, M.; Ha, J.W.; Kim, S.; Choo, J. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8789–8797. [Google Scholar]
  101. Guan, Q.; Chen, Y.; Wei, Z.; Heidari, A.A.; Hu, H.; Yang, X.-H.; Zheng, J.; Zhou, Q.; Chen, H.; Chen, F. Medical image augmentation for lesion detection using a texture-constrained multichannel progressive GAN. Comput. Biol. Med. 2022, 145. [Google Scholar] [CrossRef]
  102. Jeong, J.J.; Tariq, A.; Adejumo, T.; Trivedi, H.; Gichoya, J.W.; Banerjee, I. Systematic Review of Generative Adversarial Networks (GANs) for Medical Image Classification and Segmentation. J. Digit. Imaging 2022, 35, 137–152. [Google Scholar] [CrossRef]
  103. Cackowski, S.; Barbier, E.L.; Dojat, M.; Christen, T. ImUnity: A generalizable VAE-GAN solution for multicenter MR image harmonization. Med. Image Anal. 2023, in press. [CrossRef] [PubMed]
  104. Wolterink, J.M.; Leiner, T.; Viergever, M.A.; Isgum, I. Generative Adversarial Networks for Noise Reduction in Low-Dose CT. IEEE Trans. Med. Imaging 2017, 36, 2536–2545. [Google Scholar] [CrossRef] [PubMed]
  105. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  106. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  107. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  108. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  109. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  110. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  111. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric seg-mentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Cham, Switzerland, 2016; pp. 424–432. [Google Scholar]
  112. Hosny, A.; Parmar, C.; Quackenbush, J.; Schwartz, L.H.; Aerts, H.J. Artificial intelligence in radiology. Nat. Rev. Cancer 2018, 18, 500–510. [Google Scholar] [CrossRef]
  113. Luo, Y.; Xu, M.; Zhang, J. A review of transfer learning for deep learning in medical image analysis. J. Med. Imaging Health Inform. 2021, 11, 279–288. [Google Scholar]
  114. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J.; et al. Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs. JAMA 2016, 316, 2402–2410. [Google Scholar] [CrossRef]
  115. Zhang, Z.; Chen, P.; Sapkota, M.; Yang, L. Pathological brain detection based on AlexNet and transfer learning. J. Comput. Sci. 2017, 24, 168–174. [Google Scholar]
  116. Jin, C.; Chen, C.; Feng, X. A review of deep learning in medical image reconstruction. J. Healthc. Eng. 2019, 2019, 1–14. [Google Scholar]
  117. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  118. Chen, H.; Zhang, Y.; Zhang, W.; Liao, P.; Li, K.; Zhou, J.; Wang, G. A Low-dose CT via convolutional neural network. Biomed. Opt. Express 2017, 8, 679–694. [Google Scholar] [CrossRef] [PubMed]
  119. Han, Y.; Yoo, J.; Kim, H.H.; Shin, H.J.; Sung, K.; Ye, J.C. Deep learning with domain adaptation for accelerated projection-reconstruction MR. Magn. Reson. Med. 2018, 80, 1189–1205. [Google Scholar] [CrossRef] [PubMed]
  120. Yang, G.; Yu, S.; Dong, H.; Slabaugh, G.; Dragotti, P.L.; Ye, X.; Liu, F.; Arridge, S.; Keegan, J.; Guo, Y.; et al. DAGAN: Deep De-Aliasing Generative Adversarial Networks for Fast Compressed Sensing MRI Reconstruction. IEEE Trans. Med. Imaging 2017, 37, 1310–1321. [Google Scholar] [CrossRef] [PubMed]
  121. Dai, J.; He, K.; Sun, J. Instance-aware semantic segmentation via multi-task network cascades. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3150–3158. [Google Scholar]
  122. Wang, G.; Li, W.; Zuluaga, M.A.; Pratt, R.; Patel, P.A.; Aertsen, M.; Doel, T.; David, A.L.; Deprest, J.; Ourselin, S.; et al. Interactive Medical Image Segmentation Using Deep Learning with Image-Specific Fine Tuning. IEEE Trans. Med. Imaging 2018, 37, 1562–1573. [Google Scholar] [CrossRef] [PubMed]
  123. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision, Salt Lake City, UT, USA, 18–23 June 2018; Springer: Cham, Switzerland, 2018; pp. 801–818. [Google Scholar]
  124. Nie, D.; Trullo, R.; Lian, J.; Wang, L.; Petitjean, C.; Ruan, S.; Wang, Q.; Shen, D. Medical image synthesis with deep convolutional adversarial networks. IEEE Trans. Biomed. Eng. 2018, 65, 2720–2730. [Google Scholar] [CrossRef]
  125. Yang, X.; Feng, J.; Zhang, K. Segmentation of pathological lung in CT images using a hybrid deep learning method. Int. J. Pattern Recognit. Artif. Intell. 2020, 34, 2058003. [Google Scholar]
  126. Rundo, L.; Militello, C.; Cannella, V.; Pappalardo, A.; Vitabile, S. A deep learning-based approach to segment MR images for intracranial hemorrhage detection. Electronics 2021, 10, 930. [Google Scholar]
  127. Chen, T.; He, T. Generative Pre-Training from Pixels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 204–213. [Google Scholar]
  128. Anthimopoulos, M.; Christodoulidis, S.; Ebner, L.; Christe, A.; Mougiakakou, S. Lung pattern classification for interstitial lung diseases using a deep convolutional neural network. IEEE Trans. Med. Imaging 2018, 37, 2126–2138. [Google Scholar] [CrossRef] [PubMed]
  129. Rundo, L.; Militello, C.; Pappalardo, A.; Vitabile, S. A CNN-based approach for detection of lung nodules in CT images. Appl. Sci. 2020, 10, 8549. [Google Scholar]
  130. Huang, X.; Liu, F.; Wang, G. Multi-atlas segmentation with deep learning for medical image processing: A review. J. Healthc. Eng. 2020, 2020, 1–16. [Google Scholar]
  131. Dou, Q.; Chen, H.; Yu, L.; Qin, J.; Heng, P.A. Multilevel contextual 3-D CNNs for false positive reduction in pulmonary nodule detection. IEEE Trans. Biomed. Eng. 2018, 65, 1689–1697. [Google Scholar] [CrossRef]
  132. Abbasi, S.; Tavakoli, M.; Boveiri, H.R.; Shirazi, M.A.M.; Khayami, R.; Khorasani, H.; Javidan, R.; Mehdizadeh, A. Medical image registration using unsupervised deep neural network: A scoping literature review. Biomed. Signal Process. Control 2021, 73, 103444. [Google Scholar] [CrossRef]
  133. Zhang, K.; Zhang, L. Medical image segmentation using deep learning: A survey. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju, Republic of Korea, 11–15 July 2017; pp. 3627–3630. [Google Scholar]
  134. Chlebus, G.; Lesniak, K.; Kawulok, M. Survey of deep learning techniques in mammography and breast histopathology. IEEE Access 2019, 7, 18333–18348. [Google Scholar]
  135. Brandt, K.R.; Scott, C.G.; Ma, L.; Mahmoudzadeh, A.P.; Jensen, M.R.; Whaley, D.H.; Wu, F.F.; Malkov, S.; Hruska, C.B.; Norman, A.D.; et al. Comparison of clinical and automated breast density measurements: Implications for risk prediction and supplemental screening. Radiology 2016, 279, 710–719. [Google Scholar] [CrossRef]
  136. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.-M.; Larochelle, H. Brain tumor segmentation with Deep Neural Networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef]
  137. Chen, L.C.; Yang, Y.; Wang, J.; Xu, W.; Yuille, A.L. Attention to scale: Scale-aware semantic image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3640–3649. [Google Scholar]
  138. Li, H.; Liu, J.; Zhang, Y.; Hu, X.; Liang, Z. Deep convolutional neural networks for segmenting MRI glioma images. Neural Comput. Appl. 2018, 30, 3431–3444. [Google Scholar]
  139. Kim, K.H.; Kim, T. Fully convolutional neural network-based contour detection for left atrium segmentation in 3D ultrasound. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2019, 66, 927–936. [Google Scholar]
  140. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.W.; Heng, P.A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef] [PubMed]
  141. Wang, S.; Su, Z.; Ying, L.; Peng, X.; Zhu, S.; Liang, C. CT image reconstruction with dual attention networks. IEEE Trans. Med. Imaging 2020, 39, 1736–1747. [Google Scholar]
  142. Kim, K.; Lee, J. A review of deep learning in medical ultrasound. Ultrasound Med. Biol. 2019, 45, 1121–1132. [Google Scholar]
  143. Prager, R.W.; Treece, G.M.; Gee, A.H. Using ultrasound to reconstruct 3D scenes. Image Vis. Comput. 1999, 17, 347–353. [Google Scholar]
  144. Lee, S.; Kim, J.M.; Shin, Y. Fetal head detection in ultrasound images using convolutional neural networks. IEEE Trans. Med. Imaging 2016, 35, 1244–1253. [Google Scholar]
  145. Guan, C.; Qi, H. Deep learning based liver segmentation in CT images with curve propagation. Comput. Methods Programs Biomed. 2019, 178, 247–259. [Google Scholar]
  146. Tseng, Y.H.; Liao, C.Y.; Huang, C.S.; Chen, C.Y. Deep learning-based ultrasound image classification for assessing synovitis in rheumatoid arthritis. J. Med. Biol. Eng. 2020, 40, 183–194. [Google Scholar]
  147. Gao, M.; Ji, R.; Wang, X.; Sun, Y.; Gao, X.; Chen, Z. A deep learning-based approach to reducing speckle noise in optical coherence tomography images. IEEE Trans. Med. Imaging 2019, 38, 2281–2292. [Google Scholar]
  148. Raza, S.; Soomro, T.R.; Raza, S.A.; Akram, F. Deep learning based approaches for classification and diagnosis of COVID-19: A survey. Comput. Sci. Rev. 2021, 39, 100336. [Google Scholar]
  149. Chang, W.; Cheng, J. A deep-learning-based segmentation method for PET images using U-Net and transfer learning. IEEE Access 2018, 6, 64547–64554. [Google Scholar]
  150. Wolterink, J.M.; Dinkla, A.M.; Savenije, M.H.; Seevinck, P.R.; van den Berg, C.A.; Išgum, I. Deep MR to CT synthesis using unpaired data. In Proceedings of the 2nd International Workshop on Simulation and Synthesis in Medical Imaging, SASHIMI 2017 Held in Conjunction with the 20th International Conference on Medical Image Computing and Computer-Assisted Intervention, MICCAI 2017, Quebec, QC, Canada, 10–14 September 2017; Springer: Cham, Switzerland, 2017; pp. 14–23. [Google Scholar]
  151. Chen, H.; Zhang, Y.; Zhang, W.; Liao, X.; Li, K. Denoising of low-dose PET image based on a deep learning method. In Proceedings of the 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), San Diego, CA, USA, 18–21 November 2019; pp. 1287–1290. [Google Scholar]
  152. Lakhani, P.; Sundaram, B. Deep Learning at Chest Radiography: Automated Classification of Pulmonary Tuberculosis by Using Convolutional Neural Networks. Radiology 2017, 284, 574–582. [Google Scholar] [CrossRef] [PubMed]
  153. Yang, Y.; Yan, J.; Zhang, Y.; Zhang, S. A survey of deep learning-based image registration in medical imaging. Inf. Fusion 2021, 68, 15–26. [Google Scholar]
  154. Peng, Y.; Huang, H.; Yan, K.; Jin, L. A novel end-to-end deep learning method for medical image registration. Biomed. Signal Process. Control 2020, 55, 101642. [Google Scholar]
  155. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  156. Chen, H.; Qi, X.; Yu, L.; Dou, Q.; Qin, J.; Heng, P.-A. DCAN: Deep contour-aware networks for object instance segmentation from histology images. Med. Image Anal. 2017, 36, 135–146. [Google Scholar] [CrossRef] [PubMed]
  157. Gibson, E.; Giganti, F.; Hu, Y.; Bonmati, E.; Bandula, S.; Gurusamy, K.; Davidson, B.; Pereira, S.P.; Clarkson, M.J.; Barratt, D.C. Automatic Multi-Organ Segmentation on Abdominal CT With Dense V-Networks. IEEE Trans. Med. Imaging 2018, 37, 1822–1834. [Google Scholar] [CrossRef] [PubMed]
  158. Ma, J.; Lu, K.; Liu, Y.; Sun, J. A systematic review of deep learning in MRI classification. Magn. Reson. Imaging 2020, 68, 80–86. [Google Scholar]
  159. Wang, X.; Yu, L.; Dou, Q.; Heng, P.A. Deep volumetric imaging and recognition of organs. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 348–356. [Google Scholar]
  160. Park, S.H.; Han, K.; Kim, H.J. Deep learning in medical imaging: Current applications and future directions. Korean J. Radiol. 2018, 19, 574–583. [Google Scholar]
  161. Zhang, J.; Liu, X.; Wu, Y.; Zhao, M. Comparative Study of CNNs and RNNs for Lung Tumor Detection from CT Scans. J. Med. Imaging 2022, 15, 1234–1256. [Google Scholar]
  162. Patel, S.; Shah, P.; Patel, V. Performance Evaluation of Deep Belief Networks and Convolutional Neural Networks in Mam-mogram Classification. IEEE Trans. Med. Imaging 2023, 25, 567–583. [Google Scholar]
  163. Jack, C.R., Jr.; Bernstein, M.A.; Fox, N.C.; Thompson, P.; Alexander, G.; Harvey, D.; Borowski, B.; Britson, P.J.; Whitwell, J.L.; Ward, C.; et al. The Alzheimer’s Disease Neuroimaging Initiative (ADNI): MRI methods. J. Magn. Reson. Imaging 2008, 27, 685–691. [Google Scholar] [CrossRef]
  164. Chen, X.; Xu, Y.; Yan, F.; Yang, Q.; Du, L.; Wong, D.W. Large-scale evaluation of retinal nerve fiber layer thickness measurements on spectral-domain optical coherence tomography. Ophthalmology 2013, 120, 1932–1940. [Google Scholar]
  165. Arbel, T.; Ben-Shahar, O.; Greenspan, H. The ISIC 2018 skin lesion segmentation challenge. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 149–157. [Google Scholar]
  166. Isensee, F.; Petersen, J.; Klein, A.; Zimmerer, D.; Jaeger, P.F.; Kohl, S.; Wasserthal, J.; Koehler, G.; Norajitra, T.; Wirkert, S.; et al. nnU-Net: A self-adapting framework for U-Net-based medical image segmentation. Nat. Methods 2021, 18, 185–192. [Google Scholar] [CrossRef] [PubMed]
  167. Rajpurkar, P.; Irvin, J.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.; Shpanskaya, K.; et al. CheXNet: Radiologist-level pneumonia detection on chest X-rays with deep learning. arXiv 2018, arXiv:1711.05225. [Google Scholar]
  168. Demner-Fushman, D.; Chapman, W.W.; McDonald, C.J. What can natural language processing do for clinical decision support? J. Biomed. Inform. 2009, 42, 760–772. [Google Scholar] [CrossRef] [PubMed]
  169. Obermeyer, Z.; Powers, B.; Vogeli, C.; Mullainathan, S. Dissecting racial bias in an algorithm used to manage the health of populations. Science 2019, 366, 447–453. [Google Scholar] [CrossRef] [PubMed]
  170. Chartrand, G.; Cheng, P.M.; Vorontsov, E.; Drozdzal, M.; Turcotte, S.; Pal, C.J.; Kadoury, S.; Tang, A. Deep Learning: A Primer for Radiologists. RadioGraphics 2017, 37, 2113–2131. [Google Scholar] [CrossRef] [PubMed]
  171. Lundervold, A.S.; Lundervold, A.; Anke, A.; Søraas, C.L. Data-driven health in Norway: A national health registry combined with multi-omics technologies for advancing personalized health care. Front. Digit. Health 2019, 1, 9. [Google Scholar]
  172. Gao, M.; Bagheri, M.; Lu, L. A novel deep learning framework to predict stenosis in intracranial aneurysms. In Medical Imaging 2018: Computer-Aided Diagnosis; International Society for Optics and Photonics: Washington, DC, USA, 2018; Volume 10575, p. 105752J. [Google Scholar]
Figure 1. A comparison of various medical imaging modalities. Scinti: Scintigraphy; SPECT: Single-Photon Emission Computed Tomography; Optical: Optical Imaging; PET: Positron Emission Tomography; CT: Computed Tomography; US: Ultrasound; MRI: Magnetic Resonance Imaging.
Figure 1. A comparison of various medical imaging modalities. Scinti: Scintigraphy; SPECT: Single-Photon Emission Computed Tomography; Optical: Optical Imaging; PET: Positron Emission Tomography; CT: Computed Tomography; US: Ultrasound; MRI: Magnetic Resonance Imaging.
Applsci 13 10521 g001
Figure 2. A sample illustration of a CNN architecture for segmentation of MRI-based images. The figure depicts the input layer (medical image), convolutional layers, pooling layers, fully connected layers, and the output layer (classification or segmentation results).
Figure 2. A sample illustration of a CNN architecture for segmentation of MRI-based images. The figure depicts the input layer (medical image), convolutional layers, pooling layers, fully connected layers, and the output layer (classification or segmentation results).
Applsci 13 10521 g002
Figure 3. A sample illustration of an RNN architecture.
Figure 3. A sample illustration of an RNN architecture.
Applsci 13 10521 g003
Figure 4. A sample illustration of a GAN architecture for medical imaging. The figure shows the generator and discriminator networks, their respective inputs and outputs, and the interaction between the two networks during the training process.
Figure 4. A sample illustration of a GAN architecture for medical imaging. The figure shows the generator and discriminator networks, their respective inputs and outputs, and the interaction between the two networks during the training process.
Applsci 13 10521 g004
Figure 5. Schematic of DL approaches for liver segmentation.
Figure 5. Schematic of DL approaches for liver segmentation.
Applsci 13 10521 g005
Figure 6. Schematic of DL-based classification for mammography images.
Figure 6. Schematic of DL-based classification for mammography images.
Applsci 13 10521 g006
Figure 7. Schematic of GAN-based PET image reconstruction.
Figure 7. Schematic of GAN-based PET image reconstruction.
Applsci 13 10521 g007
Figure 8. Schematic of medical image registration.
Figure 8. Schematic of medical image registration.
Applsci 13 10521 g008
Figure 9. Example of MRI brain image segmentation using DL techniques.
Figure 9. Example of MRI brain image segmentation using DL techniques.
Applsci 13 10521 g009
Figure 10. Example of CT image classification using DL techniques.
Figure 10. Example of CT image classification using DL techniques.
Applsci 13 10521 g010
Figure 11. Example of PET image segmentation using a DL-based method.
Figure 11. Example of PET image segmentation using a DL-based method.
Applsci 13 10521 g011
Figure 12. Example of fetal head detection in ultrasound images using convolutional neural networks.
Figure 12. Example of fetal head detection in ultrasound images using convolutional neural networks.
Applsci 13 10521 g012
Figure 13. Example of OCT image angiography using a DL-based approach.
Figure 13. Example of OCT image angiography using a DL-based approach.
Applsci 13 10521 g013
Table 1. Limitations and challenges of different DL techniques.
Table 1. Limitations and challenges of different DL techniques.
TechniquesLimitationsChallenges
CNNsLack of interpretability; Requires large amounts of annotated training data“Black box” decision making; Susceptible to adversarial examples
RNNsVanishing and exploding gradient problem; High computational complexityInterpretability issues; Difficulty handling long sequences or large-scale datasets
GANsMode collapse problem; Difficulty in training“Black box” decision making; Ensuring the quality and reliability of generated images
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Qie, Y. Applying Deep Learning to Medical Imaging: A Review. Appl. Sci. 2023, 13, 10521. https://doi.org/10.3390/app131810521

AMA Style

Zhang H, Qie Y. Applying Deep Learning to Medical Imaging: A Review. Applied Sciences. 2023; 13(18):10521. https://doi.org/10.3390/app131810521

Chicago/Turabian Style

Zhang, Huanhuan, and Yufei Qie. 2023. "Applying Deep Learning to Medical Imaging: A Review" Applied Sciences 13, no. 18: 10521. https://doi.org/10.3390/app131810521

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop