Next Article in Journal
The ETS Homologous Factor (EHF) Represents a Useful Immunohistochemical Marker for Predicting Prostate Cancer Metastasis
Previous Article in Journal
Comparison of Linear and Convex-Array Transducers in Assessing the Enhancement Characteristics of Suspicious Breast Lesions at Contrast-Enhanced Ultrasound (CEUS)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Current Value of Biparametric Prostate MRI with Machine-Learning or Deep-Learning in the Detection, Grading, and Characterization of Prostate Cancer: A Systematic Review

1
Medical Faculty Mannheim, University of Heidelberg, 69120 Heidelberg, Germany
2
Academic Radiology, Department of Translational Research, University of Pisa, 56126 Pisa, Italy
3
Italian Society of Medical and Interventional Radiology, SIRM Foundation, Via della Signora 2, 20122 Milano, Italy
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(4), 799; https://doi.org/10.3390/diagnostics12040799
Submission received: 22 February 2022 / Revised: 19 March 2022 / Accepted: 23 March 2022 / Published: 24 March 2022
(This article belongs to the Topic Artificial Intelligence in Cancer Diagnosis and Therapy)

Abstract

:
Prostate cancer detection with magnetic resonance imaging is based on a standardized MRI-protocol according to the PI-RADS guidelines including morphologic imaging, diffusion weighted imaging, and perfusion. To facilitate data acquisition and analysis the contrast-enhanced perfusion is often omitted resulting in a biparametric prostate MRI protocol. The intention of this review is to analyze the current value of biparametric prostate MRI in combination with methods of machine-learning and deep learning in the detection, grading, and characterization of prostate cancer; if available a direct comparison with human radiologist performance was performed. PubMed was systematically queried and 29 appropriate studies were identified and retrieved. The data show that detection of clinically significant prostate cancer and differentiation of prostate cancer from non-cancerous tissue using machine-learning and deep learning is feasible with promising results. Some techniques of machine-learning and deep-learning currently seem to be equally good as human radiologists in terms of classification of single lesion according to the PIRADS score.

1. Introduction

1.1. Prostate Cancer

Prostate cancer (PCA) is the second most common cancer in men worldwide and it accounts for up to 25% of all malignancies in Europe [1]. It is the third leading cause of cancer-related death in the United States and Europe [2,3]. The incidence of prostate cancer increases with rising age of patients, and prostate cancer and its management are becoming a major public health challenge. PCA aggressiveness can be linked to specific genes such as BRCA, and behavior such as smoking [4,5]. Accurate and early detection of prostate cancer is therefore paramount to achieve good overall patient outcomes. The tools available for assessing and detecting prostate cancer are digital rectal examination (DRE), PSA screening, transrectal ultrasound, and MRI whereby the latter received the highest amount of attention in the past decade due to its unprecedented capabilities in accuracy [6,7,8].
In contrast to ultrasound and digital rectal examination, MRI offers an operator-independent tool for objectively assessing the entire prostate gland from base to apex and from the posterior peripheral zone (PZ) to the anterior fibromuscular stroma (AFMS) that are barely assessable with DRE [6,9].
Magnetic resonance imaging of the prostate has a long history going back more than 20 years. In the initial phase, high resolution T2-weighted (T2w) imaging and spectroscopy were mainly used as tools for detecting prostate cancer. Yet, spectroscopy is slow and susceptible to artefacts and was not well perceived. In the recent decade, further developments have taken over including diffusion weighted imaging (DWI), dynamic contrast enhanced imaging (DCE). The entire prostate exam has been standardized worldwide and its reporting has been harmonized by the PIRADS (Prostate Imaging Reporting and Data System) system [10]. This classification system allows to objectively assess the prostate and potential cancerous zones and standardizes reporting over separate sites so that the overall performance of MRI is increased and is more reproducible compared to previous periods. With this development MRI of the prostate follows the trend to standardize the entire radiological procedure from image acquisition to data reporting to achieve a higher reliability, enhanced reproducibility, and a direct implication for radiology-based treatments as it has previously successfully demonstrated in breast imaging with BIRADS (Breast Imaging Reporting and Data System) [11].
The report structuring provided by PIRADS is already a condensation of the imaging information and standardizes reporting and its output. This is one major step toward a more automated and operator-independent radiology. Moreover, the image acquisition parameters, slice orientations, and sequences with its specific sequence characteristics are governed by PIRADS [12]. This automatically sets the stage for a potential automated image analysis. In the past decade, artificial intelligence (AI) with its subdivisions of machine learning (ML), radiomics, and deep learning (DL) has become more prevalent. At this point in time, ML and DL are still no clinical standards. Radiomics, for example, use quantitative imaging features that are often unrecognizable to the human eye. Therefore, it is increasing the number of potential parameters to the multi-parametric approach of prostate MRI and with potential benefits for PCA detection and grading and beyond. DL techniques such as convoluted neural networks (CNN) are currently considered gold standard in computer vision and pattern recognition and hence have potential benefits for PCA detection and grading. With larger data sets as basis, they have the potential to automatically learn and deduct conclusions so that PCA recognition based on unperceivable features to the human eye might be possible. Despite numerous experimental studies which will be discussed further in this study, there is no standardized approach on how to use and implement DL and ML for prostate imaging now.
The aim of this study is to elucidate the status of artificial intelligence in prostate imaging with a focus on the so-called bi-parametric (bp) approach of prostate MRI (bpMRI).

1.2. Prostate Imaging Reporting and Data System

PIRADS was established by key global experts in the field of prostate imaging from America and Europe (European Society of Urogenital Radiology (ESUR), American College of Radiology (ACR)) to facilitate and standardize prostate MRI with the aim of assessing the risk of clinically significant prostate cancer (csPCA). The first version of the PIRADS recommendations was published in December 2011, the latest and current update was published in 2019 (PIRADS v2.1) [10,12,13].
Various studies have compared the predictive performance of PI-RADS v1 for the detection of csPCA compared to image-guided biopsy or radical prostatectomy (RP) specimens as standard of reference. In a 2015 study, Thompson reported multi-parametric MRI detection of csPCA had sensitivity of 96%, specificity of 36%, negative predictive value and positive predictive values of 92% and 52%; when PI-RADS was incorporated into a multivariate analysis (PSA, digital rectal exam, prostate volume, patient age) the area under the curve (AUC) improved from 0.776 to 0.879, p < 0.001 [14]. A similar paper showed that PI-RADS v2 correctly identified 94–95% of prostate cancer foci ≥ 0.5 mL but was limited for the assessment of Gleason Score (GS) ≥ 4 + 3 csPCA ≤ 0.5 mL [15]. An experienced radiologist using PIRADS v2 is reported to achieve an AUC of 0.83 with 77% sensitivity and 81% specificity [16].

1.3. Sequences for Prostate MRI

The initial protocol for MRI of the prostate as provided by PIRADS included high-resolution multiplanar T2w-imaging, DWI, and DCE after the intravenous administration of paramagnetic gadolinium chelate contrast agent. This so-called multi parametric prostate MRI (mpMRI) is considered as the gold standard. T2w-imaging is used to demonstrate zonal anatomy of the prostate. Tumors can be well delineated, and their relation to the prostate capsule can be examined. Benign changes such as benign prostate hyperplasia, post-prostatic changes of the peripheral zone or scars can be identified. T2w-imaging is considered the gold standard for the transitional zone (TZ) of the prostate gland. In addition, T2w-imaging can be used to measure the volume of the prostate. The high anatomic information content of T2w-imaging makes this sequence the perfect roadmap for image-guided biopsy [12,17].
DWI serves as an indirect measure of cellular density. In case of a malignant tumor with high cellular density, the ability of water to freely move in the interstitial compartment is decreased hence the diffusion is impaired. The images with high b-values and even those with more and more common-interpolated calculated b-values allow quick and easy depiction of these suspicious areas in the prostate. The calculated ADC maps give a quantitative measure of cellular density and can be considered as a molecular imaging tool for tumor aggressiveness. DWI imaging is considered as the reference sequence for the peripheral zone (PZ) of the prostate [12,17].
Dynamic contrast enhancement (DCE) is considered as the weakest of the three used approaches for prostate imaging. In contrast to T2w-imaging and DWI, DCE is not being considered as a dominant sequence for any of the prostate zones. It only serves as a tiebreaker in very specific questions in the PIRADS system. In addition, it requires the intravenous administration of contrast agent with the risk of side-effects such as allergies, nephrogenic systemic fibrosis, or Gadolinium deposition in the body [18,19,20,21]. While the risk of nephrogenic systemic fibrosis is controllable by using little amounts of macrocyclic Gd-chelates, no harmful consequence for Gd-chelate depositions in the body has been found [22,23]. Nevertheless, patients often try to avoid contrast agent if feasible. Moreover, physicians embrace the idea of non-enhanced exams equally, as it speeds up the acquisition and reduces the number of potential complications. In addition, omitting contrast agent permits to save money.

1.4. Multiparametric and Biparametric MRI of the Prostate

With this in mind and the knowledge that the performance of DCE often yielded limited added value to T2w-imaging and DWI in mpMRI of the prostate bi-parametric MRI (bpMRI) of the prostate is gaining considerable support [15]. Meanwhile, there are several high-ranked studies such as the PROMIS trial and meta-analyses comparing mpMRI and bpMRI of the prostate [24,25,26]. Current data underline the high negative value of bpMRI in biopsy-naïve patients with a negative predictive value of up to 97% [27,28]. Whether bpMRI might be slightly less accurate in less-experience readers is not yet clearly proven [29,30]. A currently accepted position is that bpMRI of the prostate seems to be equally good as mpMRI of the prostate for patients with low and high risk for csPCA but DCE might be of worth in patients with intermediate risk and PIRADS 3 lesions [25,26,31,32,33,34,35] (Figure 1). bpMRI of the prostate is also commonly used for computer-based postprocessing using artificial intelligence. This is due to the fact that DCE contains a fourth dimension (time) which make those images harder to algin and match with two-dimensional anatomical images such as T2w-imaging and DWI. Another drawback of DCE is that image information is not obvious. The image information on contrast media arrival and distribution which is seen as a surrogate marker for microvascular density have to be extracted using semiquantitative or quantitative pharmacokinetic models which adds another layer of complexity on postprocessing, along with the increase of time necessary to report the exams.

1.5. Artificial Intelligence (AI) for Image Postprocessing

The availability of cheap and high computing power with the additional advent of postprocessing technologies and artificial intelligence such as machine learning techniques and deep neural networks has fostered the application of those techniques for radiology tasks such as tumor detection. The current hierarchical concept of AI is depicted in Figure 2.
Machine-learning (ML) is a subfield of AI in which algorithms are trained to perform tasks by learning rules from data rather than explicit programming. Radiomics is seen as a method that extracts large numbers of features from radiological images using data characterization algorithms such as first order statistics, shape-based, histogram-based analyses, Gray Level Co-occurrence Matrix, Gray Level Run Length Matrix, Gray Level Size Zone Matrix, Gray Level Dependence Matrix, Neighboring Gray Tone Difference Matrix to name a few [36,37,38,39]. These features are said to have the potential to uncover disease characteristics that are hard to be appreciated by the naked eye. The hypothesis of radiomics is that distinctive imaging features between disease forms may be useful for detecting changes and potentially predicting prognosis and therapeutic response for various conditions such as e.g., detection of csPCA. These radiomic features are then often further analyzed using ML-techniques. An example of a radiomics ML-workflow is shown in Figure 3. An issue concerning ML-techniques is that it often requires the manual placement of a region of interest hence hereby introducing a potential source for errors and biases.
Deep learning (DL) is a subfield of AI in which algorithms are trained to perform tasks by learning patterns from data rather than explicit programming. The key factors for the increasing attention that DL attracted in the past years are the availability of large quantities of labelled data, the inexpensive and powerful computing hardware particularly graphic-processing units and improvements in training techniques and architectures. DL is a type of representation learning in which the algorithms learn a composition of features that reflect the hierarchy of structures in the data. Current state-of-the-art for medical image recognition using DL techniques are so called convoluted neural networks (CNN). These networks are characterized by an architecture of connected non-linear functions that learn multiple levels of representations of the input data thereby extracting possibly millions of features [41]. Especially CNNs in which a series of convolution of filter layers are exploited are suitable for image processing [42]. Newer techniques such as transfer learning and data augmentation, or the application of generative methods help in mitigating existing limitations of CNN [43]. The entire process of data processing within the multiple layers of a CNN with convolution filters, pooling, and maximum filtering is beyond the scope of this study. Largely simplified, one might say that bottom layers of the CNN act as a feature extractor while the top layers of the CNN act as a classifier. An overview is given in Figure 4 in which the DL workflow is compared to radiomics or the standard radiology reading process [44]. The reason that CNN-based approaches are considered superior to radiomics is that radiomics depend on hand-crafted features which is limited, whereas CNN can generate features that are most appropriate to the problem itself [45].

2. Materials and Methods

Literature research for this study took place in August 2021. A PubMed query with the search terms “prostate” and “magnetic” and “deep learning” or “machine learning” or “radiomics” was performed. The aim was to retrieve those studies which made use of ML or DL techniques to facilitate tumor detection and grading. To make sure that only current techniques were included in the analysis only publications from the year 2019 to 2021 were included. Particularly in the field of CNN the technical improvement is rapidly evolving so that elder publications might not represent the current state-of-the-art. Total of 95 publications were initially retrieved. Of these, 66 were omitted for several reasons so that 29 publications were available for analysis (see Figure 5). Clinical data (question to be answered, number of patients, age, AI-technique, lesion segmentation, MRI-technique, sensitivity, specificity, accuracy, AUC) were then manually extracted and transferred to a Microsoft Excel 365 spreadsheet (Microsoft, Redmond, WA, USA). PRISMA guidelines were followed [46]. An overview of the study according to the PRISMA guidelines can be found in the Appendix A.
This paper focuses on bpMRI. The current PIRADS guidelines state: “Given the limited role of DCE, there is growing interest in performing prostate MRI without DCE, a procedure termed “biparametric MRI” (bpMRI). A number of studies have reported data that supports the value of bpMRI for detection of csPCA in biopsy-naïve men and those with a prior negative biopsy”. The potential benefits of bpMRI include: (1) elimination of adverse events and gadolinium, (2) faster MRI-exam times, and (3) overall reduced costs [47]. These factors will potentially make bpMRI easily accessible. Remaining concerns are that the DCE sequence may serve as backup in case of image degradation of the DWI or T2w sequence. It seems as if DCE may be of less value for assessment of treatment of naïve prostate patients but remains essential in assessment for local recurrence following prior treatment, which however is a setting in which current PI-RADS assessment criteria do not apply. The conclusion of the PIRADS steering committee therefore advocates the use of mpMRI particularly in (1) patients with prior negative biopsies with unexplained raised PSA values, (2) those in active surveillance who are being evaluated for fast PSA doubling times or changing clinical/pathologic status, (3) men who previously had undergone a bpMRI exam that did not show findings suspicious for csPCA, and who remain at persistent suspicion of harboring disease, (4) biopsy-naïve men with strong family history, known genetic predispositions, elevated urinary genomic scores, and higher than average risk calculator scores for csPCA, and (5) men with a hip implant or other consideration that will likely degrade DWI [47].
For this paper bpMRI was selected as most studies dealing with ML or DL techniques solely relay on T2w-imaging and DWI. DCE data were rarely included. In contrast to T2w-imaging and DWI the DCE-data must be postprocessed first to generate parameter maps. This process is not yet standardized as several pharmacokinetic models and hereof derived software implementations for postprocessing exist. Without generation of parameter maps a huge number of images would have to be fed into the ML/DL algorithms—a step that most research groups obviously did not want to undertake.

3. Results

All included studies are listed with an abbreviated overview in Table 1.
Total of 29 studies were included in this study. Thirteen of them used ML (44.8%), 14 of them used DL-techniques (48.2%), and 2 of them used a combination of ML and DL (6.9%). The data for 27 of the studies were acquired at 3T (93.1%), 2 of them were acquired at 1.5 T (6.9%). A total of 7466 patients were analyzed within this data set. Hereby, the ProstatEx-data set from the Radbound University, The Netherlands was used seven times. The smallest study had a sample size of 25 patients, the largest study had a sample size of 834 patients. The MRI-technique used for AI-postprocessing most often was T2w-imaging in combination with ADC map and DWI (15 studies/53.6%). Runner-up were T2w-imaging and ADC map (8 studies, 28.6%) and T2w-imaging and DWI (2 studies, 7.1%).

3.1. Tumor Detection and Grading

As seen in Table 1, the results (AUC, sensitivities and specificities) were comparable and no trend clearly favoring ML or DL-approaches in terms of superiority could be detected. Most studies required manual interaction in which a radiologist had to segment the region of interest.
Overall, the rate of detection and correct tumor creating using AI-techniques was comparable to the performance of trained radiologists in most studies. Studies were often hard to compare as they differed in terms of standard of reference (e.g., Gleason score (GS) vs. PIRADS vs. National Comprehensive Cancer Network Guidelines vs. ISUP Guidelines) and different cut-off values within the same grading system (e.g., GS 7 was in one study considered intermediate grade, in most studies considered high-grade tumor). Some studies focused on the PZ only, while others accepted the entire gland as target tissue.
In a small study with 33 patients to predict IMRT response, GS prediction and PCA stage, GS prediction using T2w-radiomic models was found more predictive (mean AUC 0.739) rather than ADC models (mean AUC 0.70), while for stage prediction, ADC models had higher prediction performance (mean AUC 0.675). For T2w-radiomic models, mean AUC was obtained as 0.625 [40].
Using T2w-imaging and 12 b-values from diffusion along with Kurtosis analysis and T2 mapping for differentiation GS 3 + 3 vs. GS > 3 + 3 an AUC of 0.88 (95% CI 0.82–0.95) could be reached. This study with 72 patients was the only one to employ T2 mapping which, after all, was deemed as of little worth [51].
In a stringent ML-Radiomics study, an equally high AUC for tumor grading according to National Comprehensive Cancer Network guidelines in low-risk vs. high-risk (i.e., GS 8) was found for the PIRADS assessment as well as for the ML-approach (0.73 vs. 0.71, p > 0.05) [49]. Interestingly, the precision and recall were higher with the ML-approach compared to the PIRADS assessment (0.57 and 0.86 vs. 0.45 and 0.61). Similar results were found for the discrimination of ciPCA and csPCA of the PZA using a ML-Radiomics approach with extreme gradient boosting [62]. In this study performed on the ProstatEx dataset, an AUC of 0.816 for the detection of csPCA using bpMRI was found. Adding DCE slightly increased AUC to 0.870, though this was not statistically significant. Based on the same data set but using optimized CNNs Zong et al. [63] concluded that adding ktrans from DCE deteriorated sensitivity and specificity when compared to bpMRI alone from 100%/83% to 71%/88%. The optimal reported AUC of this study was 0.84.
Extremely good ML-radiomics results for differentiation ciPCA vs. csPCA with an AUC of 0.999 were found in a study by Chen et al. They could also show that ML-radiomics exhibited a higher efficacy in differentiation ciPCA from csPCA than PIRADS. A potential explanation for this, compared to the other studies, is that outstanding result might be the study inclusion/exclusion criteria: small lesions <5 mm and lesion not well delineable on MRI were excluded [52].
Somewhat poorer results were presented in a study by Gong et al. [61]. Their ML-radiomics approach that was built on T2w-imaging and b800-DWI images yielded an AUC of 0.787 and an accuracy of 69.9% for the discrimination between ciPCA and csPCA. Adding clinical data to the MRI-based data slightly degraded the results with an AUC of 0.780 and an accuracy of 68.1%. A potential reason for this poorer outcome might be a different set of inclusion parameters.
Zhong et al. compared the performance of DL and Deep Transfer Learning (DTL) with experienced radiologists. They found that DTL further improves DL. The DTL results were comparable to radiologist’s performance using PIRADS v2. They concluded that DTL might serve as an adjunct technique to support non-experienced radiologists [54]. Similar results found a study using a CNN-trained algorithm to automatically attribute PIRADS scores to suspicions lesions. A performance comparable to a human radiologist was described [64]. The lowest agreement was found with low PIRADS score, getting better with higher PIRADS scores. There was no statistically significant difference between the radiologist-assigned PIRADS score and the AI-assigned PIRADS score with regards to the presence of csPCA for PIRADS 3–5.
In contrast, for Gleason score prediction one study found better results for AI-based approaches than radiologists for PZ and TZ [76]. This could be particularly useful in the context of active surveillance.
A different study looking into aggressiveness prediction (GS > 8) found equal AUCs for AI and radiologists but higher precision and recall rates for AI than PIRADS mitigating the problem of inter-reader variability [49].
An uncommon approach was presented in [65]. The authors hereby combined Radiomics and DL-based on bpMRI with DCE and T2w-imaging. No ADC/DWI-images were used. In few patients they included, promising results with an AUC of 0.96–0.98 for Gleason score prediction were found. No further study used this subset of DCE and T2w-imaging.
The prospective IMPROD trial also examined if the addition of clinical data and RNA expression profiles of genes associated with prostate cancer increased the accuracy for detection of csPCA [58]. In this study the bpMRI based data yielded the highest AUC 0.92. Adding RNA-based data or clinical data neither improved the results nor yielded better results by itself.
Cao et al. developed an FocalNet to automatically detect and grade PCA (Figure 6) [72]. A similar work was presented by Schelb et al. [75] where a U-Net was trained to detect, segment, and grade PCA. In comparison with radiologists’ PIRADS assessment, the U-Net sensitivities and specificities for detection of PCA at different sensitivity levels (PIRADS 3 and PIRADS 4) were comparable.
Positive results for DL-based techniques with a larger number of patients (n = 312) were found in a DL-Study by Schelb et al. using a U-Net [57]. They reported a sensitivity/specificity for radiologists using PIRADS for detection of PIRADS lesions 3 and 4 respectively of 96%/88% and 22%/50% while the U-Net approach yielded 96%/92% and 31%/47% (p > 0.05). In their study the U-Net also autocontoured the prostate and the lesion with dice-coefficient of 0.89 (very good) and 0.35 (moderate) respectively.
A ML-approach to generate “attention boxes” for the detection of csPCA was published by Mehralivand et al. [60]. Their multicentric approach with data from five institutions showed an AUC of 0.749 for PIRADS assessment of csPCA and a statistically non-significant AUC of 0.775 for the ML-based approach. For the TZ only, the ML-approach yielded a higher sensitivity for detection of csPCA than PIRADS (61.8% vs. 50.8%, p = 0.001). Interestingly, the reading time for the ML-approach was on average 40s longer.
An uncommon approach for CNNs was published by Chen et al. [66]. They used U-Net CNNs to segment the prostate and intraprostatic lesions hereby segmenting the PZ, TZ, CZ, and AFMS. Their approach demonstrated impressive results: a Dice coefficient of 63% and a sensitivity and specificity of 74.1% and 99.9% respectively for correctly segmenting the prostatic zones and the suspicious lesion. Yet, in contrast to most other studies, no grading or discrimination of the suspected PCA lesion was performed. As a segmentation study this study was included in this review as it included segmentation of the prostate and detection of the tumor within the prostate.
In a screening study with 3T-bpMRI, Winkel et al. [67] could include and analyze 48 patients, all above 45 years. In a biopsy-correlated reading two human readers and a commercial prototype DL-algorithm were compared in terms of detection of tumor-suspicious lesions and grading according to PIRADS. The DL-approach had a sensitivity and specificity of 87% and 50%. Noteworthy, the DL-analysis required just 14 s.
Different ML-based models were tested and found to be highly accurate for the diagnosis of TZ PCA (sensitivity/specificity/AUC): 93.2%/98.4%/0.989) and their discrimination from BPH-nodules. Reproducibility of segmentation was excellent (DSC 0.84 tumors and 0.87 BPH). Subgroup analyses of TZ PCA vs. stromal BPH (AUC = 0.976) and in <15 mm lesions (AUC = 0.990) remained highly accurate [48].
DL-approach for detection of csPCA in patients under active surveillance was brought up by Arif et al. [68]. Initially 366 patients with low risk were included of which 292 were included in the final study. Sensitivities and specificities for csPCA segmentation rose with increasing tumor volume: tumor volumes > 0.03 cc sensitivity 82% 7 specificity of 43%, AUC 0.65; tumor volume > 0.1 cc sensitivity 85%, specificity of 52%, AUC 0.73. Tumor volumes > 0.5 sensitivity 94%, specificity 74%, AUC 0.89.
A total of six studies among the included studies compared DL/ML-approach to human radiologists [52,57,60,64,72,75]. Overall, due to the small number of studies and because of the different approaches the results cannot be analyzed together. What these studies had in common however was the finding, that at this point AI-based methods revealed a performance similar to that of the radiologists’. No study could either show an advantage of AI-methods of the radiologists or vice versa. An overview about the results can be seen in Table 2.

3.2. PIRADS 3 Lesions

Radiomics can detect with high accuracy csPCA in PI-RADS 3 lesions [59,77]. Hou et al. examined in a ML-Radiomics approach the ability of bpMRI to identify csPCA in PIRADS 3 lesions in a group of 253 patients with PIRADS 3 lesions in the TZ and PZ of whom 59 (22.4%) had csPCA [59]. The ML-Radiomics approach including T2w imaging, DWI and ADC had an AUC of 0.89 (95% CI 0.88–0.90) for predicting the presence of csPCA in a PIRADS 3 lesion.

3.3. Extracapsular Extension and Biochemical Recurrence

He et al. set up a large study including 459 patients who underwent 3T bpMRI before prostate biopsy and/or prostatectomy [69]. The aim of the study was first to differentiate between benign and malignant tissue second to predict extracapsular extension (ECE) of prostate tumor and third to predict positive surgical margins (PSM) after RP. Using Radiomics they developed and tested an algorithm that was able to achieve an AUC of 0.905 for the determination of benign and malignant tissue, 0.728 for the prediction of ECE, and a 0.766 for the prediction of PSM. Similarly, Hout et al. found an identical AUC of 0.728 for the prediction of ECE in a DL-based approach using different CNN-architectures [73]. Hence one can infer from the information derived from prostate imaging not only the current situation in the gland but can also predict future developments that might take place under therapy.
Biochemical recurrence (BCR) prediction based on radiomics features was examined in T2w-images only with higher prediction of BCR (C-index 0.802) than conventional scores, particularly also higher than the Gleason scoring system (C-index 0.583) [74]. This work is of particular interest as it first, was one of the few multicentric studies (three centers) with a relatively large number of patients (485) and second, demonstrated the ability of DL-based CNN to look beyond the prostate and infer predictions on the future course of the disease/patient.

4. Discussion

Prostate cancer is a growing medical condition already now being the second most common cancer in men in the western world. The detection and grading of prostate cancer are shifting more toward MRI and is demanding a higher number of MRI-studies to be performed and read. Currently, prostate MRI is considered a specialized exam and requires a highly specific experience to be performed and reported with high quality. A first step toward facilitation of mpMRI prostate acquisition, reading, and reporting was PIRADS, but surely not the last step [10,12,13]. To put it in a nutshell: prostate MRI is developing from the holy grail, and only a few radiologists were being able to read it competently to a commodity in radiology. This is one of the key drivers behind the growing demand for computer-assisted diagnostic tools, such as tumor detection and grading, to facilitate the diagnostic interpretation of prostate MRI also for less-trained radiologists. As the prostate is a densely packed organ with much more information for example as the sparsely packed lung, simple machine learning tools based on e.g., density differences cannot be successfully employed. To distinguish the different prostatic tissues, such as normal transitional and peripheral zones and malignant tissue, higher-developed machine learning tools are required, often based on radiomics or even deep learning techniques. In the papers included in this review, most approaches using either ML or DL were similar to radiologists in their performance [49,54,57,64,75]. For some specific applications, such as tumor detection in the TZ or detection of clinically significant cancers in PIRADS 3 lesions, AI-based methods might even be superior to radiologists’ performance [48,59].
These AI-based approaches should enable less well-trained radiologists to read and report prostate-MRI reports with good quality [57,75]. The literature review showed that different approaches to tumor grading and characterization either via ML or DL are capable of differentiating between cancerous and non-cancerous tissue. New approaches are even able to autonomously segment the prostate and the tumor within the gland overcoming a limitation of the elder approaches, where radiologists often had to manually segment the lesions, resulting in a highly time-consuming task [72,75]. Apart from many site-specific implementations of radiomics, ML and DL, another sign of maturation of AI-based approaches is that a first commercial tool was already presented [67]. Compared to the other algorithms, this commercial tool was trained by big data sets for the initial training. This development underlines again the trend in imaging toward commoditization of imaging and democratization of information technology enabling every radiologist to perform on a high-quality imaging.
Yet, there are some obstacles still to overcome. First, MRI is a tricky imaging tool. A major drawback of MRI is the lack of standard quantification of image intensities. Within the same image, the intensities for the same material vary as they are affected by bias field distortions and imaging acquisition parameters, not always perfectly standardized. In addition, not only do MR images taken on different scanner vary in image intensities, but the images for the same patient on the same scanner at different times may appear differently from each other due to a variety of scanner- and patient-dependent variables [45]. Therefore, the initial step in ML/DL image postprocessing is to normalize the MR intensity [45]. This process could induce errors, however. At last, also the reproducibility of CNNs varies resulting in interscan differences, though with less impact [78]. Second, most studies rely on single site source data. Multicentric studies are very rare hence making it harder to compare results of AI-based algorithms across different vendors and sequence parameters. Third, the choice of imaging sequences and their specific parameters is variable. This work focused on bpMRI of the prostate. Even though for a radiologist imaging with T2w-imaging and DWI imaging would be seen as biparametric, things look different in the world of AI-based post-postprocessing: sometimes T2w and ADC, sometimes T2w and a single high b-value, T2w and multiple b-values or T2, ADC and b-values were used (hereby neglecting uncommon outlier studies using DCE and T2 or T1 and T2). Even though DWI source date and ADC are based on the same acquisition, their information content seems different. It was observed in one study that the use of CHB-DWI led to higher specificity while the use of ADC led to highest sensitivity, making the choice of sensing modality useful for different clinical scenarios [79]. For example, maximizing specificity is important for surgery for removal of prostate where minimizing false positive rates to avoid unnecessary surgeries is required. On the other hand, for cancer screening, maximizing sensitivity may be useful to avoid missing cancerous patients [79]. A clear definition what would be considered as truly bpMRI or standards for AI-postprocessing has not been set up. Yet, there is a first European initiative on the development and standardization of AI-based tools for prostate MRI [44]. Fourth, DL-based CNNs are notorious for being a “black box” in terms of the how the decision was achieved. While this may not be entirely true—CNNs can be monitored at any level at some expense—they might never be as transparent as ML-based approaches hence scaring some physicians from using them on real patients outside studies. Moreover, here, commercialization of the techniques might be helpful as larger companies have the means and money to certify algorithms with the FDA or the EU and thus make them broadly (commercially) available.
As seven studies made use of the ProstatEx data, it is worth looking at the overall conclusions the creators of the dataset and initiators of the contest published [80]: the majority of the 71 methods submitted to the challenge (classifying prostate lesions as clinically significant or not) the majority of those methods outperformed random guessing. They conclude that automated classification of clinically significant cancer seems feasible. While in the second contest (computationally assigning lesions to a Gleason grade group) only two out 43 methods did marginally better than random guessing. The creators also conclude that more images and larger data sets with better annotations might be necessary to draw significant conclusions, which brings up again the question of means and money. Another conclusion that can be drawn when looking at the included studies is that 3 T imaging seems to be the standard. This is partly because there is substantial overlap in the source data (ProstatEx) and that, of course, studies are being conducted at University Medical Centers which most often have state-of-the-art equipment. For radiology departments in smaller hospitals or private practices having a 3 T system is less likely. Regarding how far the results of 3T e.g., DWI can be transferred to 1.5 T and how the technological improvement of 1.5 T in the field of signal reception and processing is supportive remain unclear. One might speculate that a state-of-the-art 1.5 T will yield comparable image quality to an elder 3 T system. Looking at the source data of the different studies one can roughly estimate that 30% of these were acquired on elder (>14 a) 3 T systems.
There are some unexpected studies with novel approaches to patient care that should be to highlighted. One was therapy assessment with pre- and post-IMRT T2w-imaging [40] for “delta radiomics”, using radiomic features extracted from MR images for predicting response in prostate cancer patients. While there was only one study with this specific design, extrapolating ECE or BCR has roughly the same line of thought: could not it be possible to predict for changes in the future with imaging features measured today [69,73,74]. The AUC values of these studies were unexpectedly high (0.801–0.905) as well as the number of included patients.

Limitations

This review has several limitations that need to be mentioned. First, ML and DL are extremely fast evolving techniques. Data provided in this review simply display a snapshot of the ongoing development. With the ever more powerful hardware and algorithms, future improvements seem likely. Most results are based on small feasibility studies, and larger applications of ML and DL in prostate imaging are not yet available. Whether their results match the promising initial studies remains unclear. Second, the inclusion criteria were narrow so that only 29 studies could be included. With the small sample size, different targets, and the different foci of the studies no wholistic analysis could be performed. Opening up the time window for the included studies would have led to inclusion of elder techniques potentially biasing the results.

5. Conclusions

In summary, this study investigated the current status of bpMRI of the prostate with postprocessing using ML and DL with a focus and tumor detection and grading. The presented results are very promising in terms of detection of csPCA and differentiation of prostate cancer from non-cancerous tissue. ML and DL seem to be equally good in terms of classification of single lesion according to the PIRADS score. Most approaches however rely on human interference and contouring the lesions. Only a few newer approaches automatically segment the entire gland and lesions, along with lesion grading according to PIRADS. There still exist a large variability and methods and just a few multicentric studies. No AI-postprocessing technique is considered gold standard at this time while there seems to be a trend toward CNNs. Regarding the actual MRI-sequences, most studies used T2w-imaging and either b-values from DWI or the ADC maps from DWI. The application of ML and DL to bpMRI postprocessing and the assistance in the reading process surely represent a step into the future of radiology. Currently however, these techniques remain at an experimental level and are not yet ready or available for a broader clinical application.

Author Contributions

Conceptualization, H.J.M., D.C., E.N. and G.A.; methodology, H.J.M., D.C., E.N. and G.A.; formal analysis, H.J.M. and G.A.; data curation, H.J.M. and G.A.; writing—original draft preparation, H.J.M. and G.A.; writing—review and editing, H.J.M., D.C., E.N. and G.A.; visualization, H.J.M.; supervision, H.J.M., D.C., E.N. and G.A. All authors have read and agreed to the published version of the manuscript.

Funding

The research for this paper was done within the University of Pisa Master Course in Oncologic Imaging as part of a research trial of the NAVIGATOR project (funded by Bando Salute 2018, Tuscany Region, Italy).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable as only literature research was performed.

Data Availability Statement

All data can be found in the original publications as listed above.

Acknowledgments

The research for this paper was done within the University of Pisa Master Course in Oncologic Imaging.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

ACRAmerican College of Radiology
ADCApparent Diffusion Coefficient
AFMSAnterior Fibromuscular Stroma (of the Prostate)
AIArtificial Intelligence
ASAnterior stroma (of the Prostate)
AUCArea under the Curve
BCRBiochemical Recurrence
bpbi-parametric
BPHBenign Prostate Hyperplasia
ciPCAClinically Insignificant Prostate Cancer
CNNConvoluted Neural Network
csPCAClinically Significant Prostate Cancer
CZCentral Zone (of the Prostate)
DCEDynamic Contrast-Enhanced Imaging
DLDeep-Learning
DREDigital Rectal Examination
DWIDiffusion-Weighted Imaging
ECEExtracapsular Extension
ESUREuropean Society of Urologic Radiology
GSGleason Score
HBVHigh b-Value (of DWI)
IMRTIntensity-Modulated Radiation Therapy
MLMachine-Learning
mpmulti-parametric
MRMagnetic Resonance
MRIMagnetic Resonance Imaging
nsPCANon-Significant Prostate Cancer
PZPeripheral Zone (of the Prostate)
PCAProstate Cancer
PIRADSProstate Imaging Reporting and Data System
PSMPositive Surgical Margins
RPRadical Prostatectomy
T2wT2-weighted Imaging
TSETurboSpinEcho
TZTransitional Zone (of the Prostate)
upuni-parametric

Appendix A

Table A1. Display of PRISMA items.
Table A1. Display of PRISMA items.
PRISMA ItemDescription
TitleCurrent Value of Biparametric Prostate MRI with Machine-Learning or Deep-Learning in the Detection, Grading and Characterization of Prostate Cancer: a systematic review.
Main objectiveAssessing the current value of deep-learning and machine-learning applied to biparametric MRI of the prostate
Inclusion and exclusion criteriaInclusion criteria:
-
Study listed in Pubmed
-
Search terms: “prostate” and “magnetic” and either “deep learning” or “machine learning” or “radiomics”
-
Full text access available through University of Heidelberg
-
Paper type: original investigation/research
-
Focus: Detection or grading of prostate cancer with biparametric prostate MRI
-
Language: English or German
-
Year of publication 2019–2021
Exclusion criteria:
-
No full text access
-
Wrong paper type: reviews, meta-analysis
-
Wrong focus (e.g., prostate segmentation, radiation therapy planning)
-
Wrong technique (uniparametric or multiparametric prostate MRI)
Information source and access timePubMed query in August 2021
Methods to assess risk of bias in included studiesNo structured program was used to assess bias in study selection. Internal review by the authors and critical appraisal of the data was performed.
Methods to present and synthesize resultsDescriptive statistics, listing in tabular form
Number of studies and participants included29 publications included
7466 participants included
Main outcomesVery heterogenous data did not allow for a general interpretation of all studies.
Tumor detection and grading with machine-learning and deep-learning techniques is feasible in trials and shows promising results. Reported values for AUC ranging from 0.71 to 0.999. In studies comparing human radiologists to deep-learning algorithms comparable, statistically not different results for tumor detection were found.
Limitations
-
No overall statistical analysis feasible due to the heterogeneity of methods and inclusion criteria reported
-
7 out of 29 studies based on the same dataset (ProstatEx, Radbound Nijmwegen, The Netherlands)
-
Heterogenous studies with different inclusion criteria and ground truth (i.e., if Gleason Grade constitutes high-grade cancer or not)
-
Often lacking demographic and statistical data
General interpretationDetection of clinically significant prostate cancer and differentiation of prostate cancer from non-cancerous tissue using machine-learning and deep learning is feasible with promising results. Some techniques of machine-learning and deep-learning currently seem to be equally good as human radiologists in terms of classification of single lesions according to the PIRADS score.
Primary source for fundingNo general funding. Publication costs are covered by the Universtiy of Pisa, Pisa, Italy.
Register name and registration numberNo registration

References

  1. Siegel, R.; Naishadham, D.; Jemal, A. Cancer statistics, 2013. CA Cancer J. Clin. 2013, 63, 11–30. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Jemal, A.; Center, M.M.; DeSantis, C.; Ward, E.M. Global patterns of cancer incidence and mortality rates and trends. Cancer Epidemiol. Biomark. Prev. 2010, 19, 1893–1907. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Ferlay, J.; Autier, P.; Boniol, M.; Heanue, M.; Colombet, M.; Boyle, P. Estimates of the cancer incidence and mortality in Europe in 2006. Ann. Oncol. 2007, 18, 581–592. [Google Scholar] [CrossRef] [PubMed]
  4. Crocetto, F.; Barone, B.; Caputo, V.F.; Fontana, M.; de Cobelli, O.; Ferro, M. BRCA Germline Mutations in Prostate Cancer: The Future Is Tailored. Diagnostics 2021, 11, 908. [Google Scholar] [CrossRef] [PubMed]
  5. Brookman-May, S.D.; Campi, R.; Henríquez, J.D.S.; Klatte, T.; Langenhuijsen, J.F.; Brausi, M.; Linares-Espinós, E.; Volpe, A.; Marszalek, M.; Akdogan, B.; et al. Latest Evidence on the Impact of Smoking, Sports, and Sexual Activity as Modifiable Lifestyle Risk Factors for Prostate Cancer Incidence, Recurrence, and Progression: A Systematic Review of the Literature by the European Association of Urology Section of Oncological Urology (ESOU). Eur. Urol. Focus. 2019, 5, 756–787. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Drudi, F.M.; Cantisani, V.; Angelini, F.; Ciccariello, M.; Messineo, D.; Ettorre, E.; Liberatore, M.; Scialpi, M. Multiparametric MRI Versus Multiparametric US in the Detection of Prostate Cancer. Anticancer Res. 2019, 39, 3101–3110. [Google Scholar] [CrossRef] [PubMed]
  7. Jones, D.; Friend, C.; Dreher, A.; Allgar, V.; Macleod, U. The diagnostic test accuracy of rectal examination for prostate cancer diagnosis in symptomatic patients: A systematic review. BMC Fam. Pract. 2018, 19, 79. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Naji, L.; Randhawa, H.; Sohani, Z.; Dennis, B.; Lautenbach, D.; Kavanagh, O.; Bawor, M.; Banfield, L.; Profetto, J. Digital Rectal Examination for Prostate Cancer Screening in Primary Care: A Systematic Review and Meta-Analysis. Ann. Fam. Med. 2018, 16, 149–154. [Google Scholar] [CrossRef] [PubMed]
  9. Pokorny, M.R.; de Rooij, M.; Duncan, E.; Schröder, F.H.; Parkinson, R.; Barentsz, J.O.; Thompson, L.C. Prospective study of diagnostic accuracy comparing prostate cancer detection by transrectal ultrasound-guided biopsy versus magnetic resonance (MR) imaging with subsequent MR-guided biopsy in men without previous prostate biopsies. Eur. Urol. 2014, 66, 22–29. [Google Scholar] [CrossRef]
  10. Barentsz, J.O.; Richenberg, J.; Clements, R.; Choyke, P.; Verma, S.; Villeirs, G.; Rouviere, O.; Logager, V.; Fütterer, J.J. ESUR prostate MR guidelines 2012. Eur. Radiol. 2012, 22, 746–757. [Google Scholar] [CrossRef] [Green Version]
  11. Spak, D.A.; Plaxco, J.S.; Santiago, L.; Dryden, M.J.; Dogan, B.E. BI-RADS® fifth edition: A summary of changes. Diagn. Interv. Imaging 2017, 98, 179–190. [Google Scholar] [CrossRef]
  12. Turkbey, B.; Rosenkrantz, A.B.; Haider, M.A.; Padhani, A.R.; Villeirs, G.; Macura, K.J.; Tempany, C.M.; Choyke, P.L.; Cornud, F.; Margolis, D.J.; et al. Prostate Imaging Reporting and Data System Version 2.1: 2019 Update of Prostate Imaging Reporting and Data System Version 2. Eur. Urol. 2019, 76, 340–351. [Google Scholar] [CrossRef]
  13. Weinreb, J.C.; Barentsz, J.O.; Choyke, P.L.; Cornud, F.; Haider, M.A.; Macura, K.J.; Margolis, D.; Schnall, M.D.; Shtern, F.; Tempany, C.M.; et al. PI-RADS Prostate Imaging-Reporting and Data System: 2015, Version 2. Eur. Urol. 2016, 69, 16–40. [Google Scholar] [CrossRef]
  14. Thompson, J.E.; van Leeuwen, P.J.; Moses, D.; Shnier, R.; Brenner, P.; Delprado, W.; Pulbrook, M.; Böhm, M.; Haynes, A.M.; Hayen, A.; et al. The Diagnostic Performance of Multiparametric Magnetic Resonance Imaging to Detect Significant Prostate Cancer. J. Urol. 2016, 195, 1428–1435. [Google Scholar] [CrossRef]
  15. Vargas, H.A.; Hötker, A.M.; Goldman, D.A.; Moskowitz, C.S.; Gondo, T.; Matsumoto, K.; Ehdaie, B.; Woo, S.; Fine, S.W.; Reuter, V.E.; et al. Updated prostate imaging reporting and data system (PIRADS v2) recommendations for the detection of clinically significant prostate cancer using multiparametric MRI: Critical evaluation using whole-mount pathology as standard of reference. Eur. Radiol. 2016, 26, 1606–1612. [Google Scholar] [CrossRef] [Green Version]
  16. Kasel-Seibert, M.; Lehmann, T.; Aschenbach, R.; Guettler, F.V.; Abubrig, M.; Grimm, M.O.; Teichgraeber, U.; Franiel, T. Assessment of PI-RADS v2 for the Detection of Prostate Cancer. Eur. J. Radiol. 2016, 85, 726–731. [Google Scholar] [CrossRef] [Green Version]
  17. Palumbo, P.; Manetta, R.; Izzo, A.; Bruno, F.; Arrigoni, F.; de Filippo, M.; Splendiani, A.; di Cesare, E.; Masciocchi, C.; Barile, A. Biparametric (bp) and multiparametric (mp) magnetic resonance imaging (MRI) approach to prostate cancer disease: A narrative review of current debate on dynamic contrast enhancement. Gland. Surg. 2020, 9, 2235–2247. [Google Scholar] [CrossRef]
  18. Michaely, H.J.; Thomsen, H.S.; Reiser, M.F.; Schoenberg, S.O. Nephrogenic systemic fibrosis (NSF)—implications for radiology. Radiologe 2007, 47, 785–793. [Google Scholar] [CrossRef]
  19. Grobner, T. Gadolinium--a specific trigger for the development of nephrogenic fibrosing dermopathy and nephrogenic systemic fibrosis? Nephrol. Dial. Transplant. 2006, 21, 1104–1108. [Google Scholar] [CrossRef] [Green Version]
  20. Alkhunizi, S.M.; Fakhoury, M.; Abou-Kheir, W.; Lawand, N. Gadolinium Retention in the Central and Peripheral Nervous System: Implications for Pain, Cognition, and Neurogenesis. Radiology 2020, 297, 407–416. [Google Scholar] [CrossRef]
  21. Radbruch, A.; Weberling, L.D.; Kieslich, P.J.; Eidel, O.; Burth, S.; Kickingereder, P.; Heiland, S.; Wick, W.; Schlemmer, H.P.; Bendszus, M. Gadolinium retention in the dentate nucleus and globus pallidus is dependent on the class of contrast agent. Radiology 2015, 275, 783–791. [Google Scholar] [CrossRef]
  22. Strickler, S.E.; Clark, K.R. Gadolinium Deposition: A Study Review. Radiol. Technol. 2021, 92, 249–258. [Google Scholar]
  23. Semelka, R.C.; Ramalho, J.; Vakharia, A.; AlObaidy, M.; Burke, L.M.; Jay, M.; Ramalho, M. Gadolinium deposition disease: Initial description of a disease that has been around for a while. Magn. Reson. Imaging 2016, 34, 1383–1390. [Google Scholar] [CrossRef] [Green Version]
  24. Ahmed, H.U.; El-Shater Bosaily, A.; Brown, L.C.; Gabe, R.; Kaplan, R.; Parmar, M.K.; Collaco-Moraes, Y.; Ward, K.; Hindley, R.G.; Freeman, A.; et al. Diagnostic accuracy of multi-parametric MRI and TRUS biopsy in prostate cancer (PROMIS): A paired validating confirmatory study. Lancet 2017, 389, 815–822. [Google Scholar] [CrossRef] [Green Version]
  25. Alabousi, M.; Salameh, J.P.; Gusenbauer, K.; Samoilov, L.; Jafri, A.; Yu, H.; Alabousi, A. Biparametric vs multiparametric prostate magnetic resonance imaging for the detection of prostate cancer in treatment-naïve patients: A diagnostic test accuracy systematic review and meta-analysis. BJU Int. 2019, 124, 209–220. [Google Scholar] [CrossRef] [PubMed]
  26. Bass, E.J.; Pantovic, A.; Connor, M.; Gabe, R.; Padhani, A.R.; Rockall, A.; Sokhi, H.; Tam, H.; Winkler, M.; Ahmed, H.U. A systematic review and meta-analysis of the diagnostic accuracy of biparametric prostate MRI for prostate cancer in men at risk. Prostate Cancer Prostatic Dis. 2020, 24, 596–611. [Google Scholar] [CrossRef] [PubMed]
  27. Boesen, L.; Nørgaard, N.; Løgager, V.; Balslev, I.; Bisbjerg, R.; Thestrup, K.C.; Winther, M.D.; Jakobsen, H.; Thomsen, H.S. Assessment of the Diagnostic Accuracy of Biparametric Magnetic Resonance Imaging for Prostate Cancer in Biopsy-Naive Men: The Biparametric MRI for Detection of Prostate Cancer (BIDOC) Study. JAMA Netw. Open 2018, 1, e180219. [Google Scholar] [CrossRef] [Green Version]
  28. Jambor, I.; Verho, J.; Ettala, O.; Knaapila, J.; Taimen, P.; Syvänen, K.T.; Kiviniemi, A.; Kähkönen, E.; Perez, I.M.; Seppänen, M.; et al. Validation of IMPROD biparametric MRI in men with clinically suspected prostate cancer: A prospective multi-institutional trial. PLoS Med. 2019, 16, e1002813. [Google Scholar] [CrossRef] [Green Version]
  29. Di Campli, E.; Delli Pizzi, A.; Seccia, B.; Cianci, R.; d’Annibale, M.; Colasante, A.; Cinalli, S.; Castellan, P.; Navarra, R.; Iantorno, R.; et al. Diagnostic accuracy of biparametric vs multiparametric MRI in clinically significant prostate cancer: Comparison between readers with different experience. Eur. J. Radiol. 2018, 101, 17–23. [Google Scholar] [CrossRef] [PubMed]
  30. Gatti, M.; Faletti, R.; Calleris, G.; Giglio, J.; Berzovini, C.; Gentile, F.; Marra, G.; Misischi, F.; Molinaro, L.; Bergamasco, L.; et al. Prostate cancer detection with biparametric magnetic resonance imaging (bpMRI) by readers with different experience: Performance and comparison with multiparametric (mpMRI). Abdom. Radiol. 2019, 44, 1883–1893. [Google Scholar] [CrossRef]
  31. Cho, J.; Ahn, H.; Hwang, S.I.; Lee, H.J.; Choe, G.; Byun, S.S.; Hong, S.K. Biparametric versus multiparametric magnetic resonance imaging of the prostate: Detection of clinically significant cancer in a perfect match group. Prostate Int. 2020, 8, 146–151. [Google Scholar] [CrossRef]
  32. Lee, D.H.; Nam, J.K.; Lee, S.S.; Han, J.Y.; Lee, J.W.; Chung, M.K.; Park, S.W. Comparison of Multiparametric and Biparametric MRI in First Round Cognitive Targeted Prostate Biopsy in Patients with PSA Levels under 10 ng/mL. Yonsei Med. J. 2017, 58, 994–999. [Google Scholar] [CrossRef] [Green Version]
  33. Woo, S.; Suh, C.H.; Kim, S.Y.; Cho, J.Y.; Kim, S.H.; Moon, M.H. Head-to-Head Comparison Between Biparametric and Multiparametric MRI for the Diagnosis of Prostate Cancer: A Systematic Review and Meta-Analysis. AJR Am. J. Roentgenol. 2018, 211, W226–W241. [Google Scholar] [CrossRef]
  34. Scialpi, M.; D’Andrea, A.; Martorana, E.; Malaspina, C.M.; Aisa, M.C.; Napoletano, M.; Orlandi, E.; Rondoni, V.; Scialpi, P.; Pacchiarini, D.; et al. Biparametric MRI of the prostate. Turk. J. Urol. 2017, 43, 401–409. [Google Scholar] [CrossRef] [Green Version]
  35. Scialpi, M.; Prosperi, E.; D’Andrea, A.; Martorana, E.; Malaspina, C.; Palumbo, B.; Orlandi, A.; Falcone, G.; Milizia, M.; Mearini, L.; et al. Biparametric versus Multiparametric MRI with Non-endorectal Coil at 3T in the Detection and Localization of Prostate Cancer. Anticancer Res. 2017, 37, 1263–1271. [Google Scholar] [CrossRef] [Green Version]
  36. Bernatz, S.; Ackermann, J.; Mandel, P.; Kaltenbach, B.; Zhdanovich, Y.; Harter, P.N.; Döring, C.; Hammerstingl, R.; Bodelle, B.; Smith, K.; et al. Comparison of machine learning algorithms to predict clinically significant prostate cancer of the peripheral zone with multiparametric MRI using clinical assessment categories and radiomic features. Eur. Radiol. 2020, 30, 6757–6769. [Google Scholar] [CrossRef]
  37. Kumar, V.; Gu, Y.; Basu, S.; Berglund, A.; Eschrich, S.A.; Schabath, M.B.; Forster, K.; Aerts, H.J.; Dekker, A.; Fenstermacher, D.; et al. Radiomics: The process and the challenges. Magn. Reson. Imaging 2012, 30, 1234–1248. [Google Scholar] [CrossRef] [Green Version]
  38. Lambin, P.; Rios-Velazquez, E.; Leijenaar, R.; Carvalho, S.; van Stiphout, R.G.; Granton, P.; Zegers, C.M.; Gillies, R.; Boellard, R.; Dekker, A.; et al. Radiomics: Extracting more information from medical images using advanced feature analysis. Eur. J. Cancer 2012, 48, 441–446. [Google Scholar] [CrossRef] [Green Version]
  39. Gillies, R.J.; Kinahan, P.E.; Hricak, H. Radiomics: Images Are More than Pictures, They Are Data. Radiology 2016, 278, 563–577. [Google Scholar] [CrossRef] [Green Version]
  40. Abdollahi, H.; Mofid, B.; Shiri, I.; Razzaghdoust, A.; Saadipoor, A.; Mahdavi, A.; Galandooz, H.M.; Mahdavi, S.R. Machine learning-based radiomic models to predict intensity-modulated radiation therapy response, Gleason score and stage in prostate cancer. Radiol. Med. 2019, 124, 555–567. [Google Scholar] [CrossRef]
  41. Wildeboer, R.R.; van Sloun, R.J.G.; Wijkstra, H.; Mischi, M. Artificial intelligence in multiparametric prostate cancer imaging with focus on deep-learning methods. Comput. Methods Programs Biomed. 2020, 189, 105316. [Google Scholar] [CrossRef]
  42. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  43. Ching, T.; Himmelstein, D.S.; Beaulieu-Jones, B.K.; Kalinin, A.A.; Do, B.T.; Way, G.P.; Ferrero, E.; Agapow, P.M.; Zietz, M.; Hoffman, M.M.; et al. Opportunities and obstacles for deep learning in biology and medicine. J. R. Soc. Interface 2018, 15, 20170387. [Google Scholar] [CrossRef] [Green Version]
  44. Penzkofer, T.; Padhani, A.R.; Turkbey, B.; Haider, M.A.; Huisman, H.; Walz, J.; Salomon, G.; Schoots, I.G.; Richenberg, J.; Villeirs, G.; et al. ESUR/ESUI position paper: Developing artificial intelligence for precision diagnosis of prostate cancer using magnetic resonance imaging. Eur. Radiol. 2021, 31, 9567–9578. [Google Scholar] [CrossRef]
  45. Chen, Q.; Hu, S.; Long, P.; Lu, F.; Shi, Y.; Li, Y. A Transfer Learning Approach for Malignant Prostate Lesion Detection on Multiparametric MRI. Technol. Cancer Res. Treat. 2019, 18, 1–9. [Google Scholar] [CrossRef]
  46. Page, M.J.; Moher, D.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. PRISMA 2020 explanation and elaboration: Updated guidance and exemplars for reporting systematic reviews. BMJ 2021, 372, n160. [Google Scholar] [CrossRef]
  47. American College of Radiology (ACR). PI-RADS. Available online: https://www.acr.org/-/media/ACR/Files/RADS/Pi-RADS/PIRADS-V2-1.pdf (accessed on 4 September 2021).
  48. Wu, M.; Krishna, S.; Thornhill, R.E.; Flood, T.A.; McInnes, M.D.F.; Schieda, N. Transition zone prostate cancer: Logistic regression and machine-learning models of quantitative ADC, shape and texture features are highly accurate for diagnosis. J. Magn. Reson. Imaging 2019, 50, 940–950. [Google Scholar] [CrossRef]
  49. Varghese, B.; Chen, F.; Hwang, D.; Palmer, S.L.; de Castro Abreu, A.L.; Ukimura, O.; Aron, M.; Aron, M.; Gill, I.; Duddalwar, V.; et al. Objective risk stratification of prostate cancer using machine learning and radiomics applied to multiparametric magnetic resonance images. Sci. Rep. 2019, 9, 1570. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Min, X.; Li, M.; Dong, D.; Feng, Z.; Zhang, P.; Ke, Z.; You, H.; Han, F.; Ma, H.; Tian, J.; et al. Multi-parametric MRI-based radiomics signature for discriminating between clinically significant and insignificant prostate cancer: Cross-validation of a machine learning method. Eur. J. Radiol. 2019, 115, 16–21. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Toivonen, J.; Montoya Perez, I.; Movahedi, P.; Merisaari, H.; Pesola, M.; Taimen, P.; Boström, P.J.; Pohjankukka, J.; Kiviniemi, A.; Pahikkala, T.; et al. Radiomics and machine learning of multisequence multiparametric prostate MRI: Towards improved non-invasive prostate cancer characterization. PLoS ONE 2019, 14, e0217702. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Chen, T.; Li, M.; Gu, Y.; Zhang, Y.; Yang, S.; Wei, C.; Wu, J.; Li, X.; Zhao, W.; Shen, J. Prostate Cancer Differentiation and Aggressiveness: Assessment With a Radiomic-Based Model vs. PI-RADS v2. J. Magn. Reson. Imaging 2019, 49, 875–884. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Xu, M.; Fang, M.; Zou, J.; Yang, S.; Yu, D.; Zhong, L.; Hu, C.; Zang, Y.; Dong, D.; Tian, J.; et al. Using biparametric MRI radiomics signature to differentiate between benign and malignant prostate lesions. Eur. J. Radiol. 2019, 114, 38–44. [Google Scholar] [CrossRef] [PubMed]
  54. Zhong, X.; Cao, R.; Shakeri, S.; Scalzo, F.; Lee, Y.; Enzmann, D.R.; Wu, H.H.; Raman, S.S.; Sung, K. Deep transfer learning-based prostate cancer classification using 3 Tesla multi-parametric MRI. Abdom. Radiol. 2019, 44, 2030–2039. [Google Scholar] [CrossRef] [PubMed]
  55. Yuan, Y.; Qin, W.; Buyyounouski, M.; Ibragimov, B.; Hancock, S.; Han, B.; Xing, L. Prostate cancer classification with multiparametric MRI transfer learning model. Med. Phys. 2019, 46, 756–765. [Google Scholar] [CrossRef] [PubMed]
  56. Xu, H.; Baxter, J.S.H.; Akin, O.; Cantor-Rivera, D. Prostate cancer detection using residual networks. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1647–1650. [Google Scholar] [CrossRef]
  57. Schelb, P.; Kohl, S.; Radtke, J.P.; Wiesenfarth, M.; Kickingereder, P.; Bickelhaupt, S.; Kuder, T.A.; Stenzinger, A.; Hohenfellner, M.; Schlemmer, H.P.; et al. Classification of Cancer at Prostate MRI: Deep Learning versus Clinical PI-RADS Assessment. Radiology 2019, 293, 607–617. [Google Scholar] [CrossRef]
  58. Montoya Perez, I.; Jambor, I.; Pahikkala, T.; Airola, A.; Merisaari, H.; Saunavaara, J.; Alinezhad, S.; Väänänen, R.M.; Tallgrén, T.; Verho, J.; et al. Prostate Cancer Risk Stratification in Men With a Clinical Suspicion of Prostate Cancer Using a Unique Biparametric MRI and Expression of 11 Genes in Apparently Benign Tissue: Evaluation Using Machine-Learning Techniques. J. Magn. Reson. Imaging 2020, 51, 1540–1553. [Google Scholar] [CrossRef]
  59. Hou, Y.; Bao, M.L.; Wu, C.J.; Zhang, J.; Zhang, Y.D.; Shi, H.B. A radiomics machine learning-based redefining score robustly identifies clinically significant prostate cancer in equivocal PI-RADS score 3 lesions. Abdom. Radiol. 2020, 45, 4223–4234. [Google Scholar] [CrossRef]
  60. Mehralivand, S.; Harmon, S.A.; Shih, J.H.; Smith, C.P.; Lay, N.; Argun, B.; Bednarova, S.; Baroni, R.H.; Canda, A.E.; Ercan, K.; et al. Multicenter Multireader Evaluation of an Artificial Intelligence-Based Attention Mapping System for the Detection of Prostate Cancer With Multiparametric MRI. AJR Am. J. Roentgenol. 2020, 215, 903–912. [Google Scholar] [CrossRef]
  61. Gong, L.; Xu, M.; Fang, M.; Zou, J.; Yang, S.; Yu, X.; Xu, D.; Zhou, L.; Li, H.; He, B.; et al. Noninvasive Prediction of High-Grade Prostate Cancer via Biparametric MRI Radiomics. J. Magn. Reson. Imaging 2020, 52, 1102–1109. [Google Scholar] [CrossRef]
  62. Bleker, J.; Kwee, T.C.; Dierckx, R.; de Jong, I.J.; Huisman, H.; Yakar, D. Multiparametric MRI and auto-fixed volume of interest-based radiomics signature for clinically significant peripheral zone prostate cancer. Eur. Radiol. 2020, 30, 1313–1324. [Google Scholar] [CrossRef] [Green Version]
  63. Zong, W.; Lee, J.K.; Liu, C.; Carver, E.N.; Feldman, A.M.; Janic, B.; Elshaikh, M.A.; Pantelic, M.V.; Hearshen, D.; Chetty, I.J.; et al. A deep dive into understanding tumor foci classification using multiparametric MRI based on convolutional neural network. Med. Phys. 2020, 47, 4077–4086. [Google Scholar] [CrossRef]
  64. Sanford, T.; Harmon, S.A.; Turkbey, E.B.; Kesani, D.; Tuncer, S.; Madariaga, M.; Yang, C.; Sackett, J.; Mehralivand, S.; Yan, P.; et al. Deep-Learning-Based Artificial Intelligence for PI-RADS Classification to Assist Multiparametric Prostate MRI Interpretation: A Development Study. J. Magn. Reson. Imaging 2020, 52, 1499–1507. [Google Scholar] [CrossRef]
  65. Brunese, L.; Mercaldo, F.; Reginelli, A.; Santone, A. Radiomics for Gleason Score Detection through Deep Learning. Sensors 2020, 20, 5411. [Google Scholar] [CrossRef]
  66. Chen, Y.; Xing, L.; Yu, L.; Bagshaw, H.P.; Buyyounouski, M.K.; Han, B. Automatic intraprostatic lesion segmentation in multiparametric magnetic resonance images with proposed multiple branch UNet. Med. Phys. 2020, 47, 6421–6429. [Google Scholar] [CrossRef]
  67. Winkel, D.J.; Wetterauer, C.; Matthias, M.O.; Lou, B.; Shi, B.; Kamen, A.; Comaniciu, D.; Seifert, H.H.; Rentsch, C.A.; Boll, D.T. Autonomous Detection and Classification of PI-RADS Lesions in an MRI Screening Population Incorporating Multicenter-Labeled Deep Learning and Biparametric Imaging: Proof of Concept. Diagnostics 2020, 10, 951. [Google Scholar] [CrossRef]
  68. Arif, M.; Schoots, I.G.; Castillo Tovar, J.; Bangma, C.H.; Krestin, G.P.; Roobol, M.J.; Niessen, W.; Veenland, J.F. Clinically significant prostate cancer detection and segmentation in low-risk patients using a convolutional neural network on multi-parametric MRI. Eur. Radiol. 2020, 30, 6582–6592. [Google Scholar] [CrossRef]
  69. He, D.; Wang, X.; Fu, C.; Wei, X.; Bao, J.; Ji, X.; Bai, H.; Xia, W.; Gao, X.; Huang, Y.; et al. MRI-based radiomics models to assess prostate cancer, extracapsular extension and positive surgical margins. Cancer Imaging 2021, 21, 46. [Google Scholar] [CrossRef]
  70. Vente, C.; Vos, P.; Hosseinzadeh, M.; Pluim, J.; Veta, M. Deep Learning Regression for Prostate Cancer Detection and Grading in Bi-Parametric MRI. IEEE Trans. Biomed. Eng. 2021, 68, 374–383. [Google Scholar] [CrossRef]
  71. Chen, J.; Wan, Z.; Zhang, J.; Li, W.; Chen, Y.; Li, Y.; Duan, Y. Medical image segmentation and reconstruction of prostate tumor based on 3D AlexNet. Comput Methods Programs Biomed. 2021, 200, 105878. [Google Scholar] [CrossRef]
  72. Cao, R.; Zhong, X.; Afshari, S.; Felker, E.; Suvannarerg, V.; Tubtawee, T.; Vangala, S.; Scalzo, F.; Raman, S.; Sung, K. Performance of Deep Learning and Genitourinary Radiologists in Detection of Prostate Cancer Using 3-T Multiparametric Magnetic Resonance Imaging. J. Magn. Reson. Imaging 2021, 54, 474–483. [Google Scholar] [CrossRef]
  73. Hou, Y.; Zhang, Y.H.; Bao, J.; Bao, M.L.; Yang, G.; Shi, H.B.; Song, Y.; Zhang, Y.D. Artificial intelligence is a promising prospect for the detection of prostate cancer extracapsular extension with mpMRI: A two-center comparative study. Eur. J. Nucl. Med. Mol. Imaging 2021, 48, 3805–3816. [Google Scholar] [CrossRef]
  74. Yan, Y.; Shao, L.; Liu, Z.; He, W.; Yang, G.; Liu, J.; Xia, H.; Zhang, Y.; Chen, H.; Liu, C.; et al. Deep Learning with Quantitative Features of Magnetic Resonance Images to Predict Biochemical Recurrence of Radical Prostatectomy: A Multi-Center Study. Cancers 2021, 13, 3098. [Google Scholar] [CrossRef]
  75. Schelb, P.; Wang, X.; Radtke, J.P.; Wiesenfarth, M.; Kickingereder, P.; Stenzinger, A.; Hohenfellner, M.; Schlemmer, H.P.; Maier-Hein, K.H.; Bonekamp, D. Simulated clinical deployment of fully automatic deep learning for clinical prostate MRI assessment. Eur. Radiol. 2021, 31, 302–313. [Google Scholar] [CrossRef]
  76. Antonelli, M.; Johnston, E.W.; Dikaios, N.; Cheung, K.K.; Sidhu, H.S.; Appayya, M.B.; Giganti, F.; Simmons, L.A.M.; Freeman, A.; Allen, C.; et al. Machine learning classifiers can predict Gleason pattern 4 prostate cancer with greater accuracy than experienced radiologists. Eur. Radiol. 2019, 29, 4754–4764. [Google Scholar] [CrossRef] [Green Version]
  77. Kan, Y.; Zhang, Q.; Hao, J.; Wang, W.; Zhuang, J.; Gao, J.; Huang, H.; Liang, J.; Marra, G.; Calleris, G.; et al. Clinico-radiological characteristic-based machine learning in reducing unnecessary prostate biopsies of PI-RADS 3 lesions with dual validation. Eur. Radiol. 2020, 30, 6274–6284. [Google Scholar] [CrossRef]
  78. Hiremath, A.; Shiradkar, R.; Merisaari, H.; Prasanna, P.; Ettala, O.; Taimen, P.; Aronen, H.J.; Boström, P.J.; Jambor, I.; Madabhushi, A. Test-retest repeatability of a deep learning architecture in detecting and segmenting clinically significant prostate cancer on apparent diffusion coefficient (ADC) maps. Eur. Radiol. 2021, 31, 379–391. [Google Scholar] [CrossRef]
  79. Dulhanty, C.; Wang, L.; Cheng, M.; Gunraj, H.; Khalvati, F.; Haider, M.A.; Wong, A. Radiomics Driven Diffusion Weighted Imaging Sensing Strategies for Zone-Level Prostate Cancer Sensing. Sensors 2020, 20, 1539. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  80. The International Society for Optics and Photonics. Solve for X: Lessons Learned from PROSTATEx. Available online: https://spie.org/news/spie-professional-magazine-archive/2019-january/solve-for-x?SSO=1 (accessed on 10 October 2021).
Figure 1. Overview of the performance of mpMRI and bpMRI based on data from Woo et al. [33] and Alabousi et al. [25] demonstrating the near equal performance of bpMRI to mpMRI (reprinted with permission from [17], Copyright 2020 Gland Surgery).
Figure 1. Overview of the performance of mpMRI and bpMRI based on data from Woo et al. [33] and Alabousi et al. [25] demonstrating the near equal performance of bpMRI to mpMRI (reprinted with permission from [17], Copyright 2020 Gland Surgery).
Diagnostics 12 00799 g001
Figure 2. Hierarchical structure of AI-techniques. Whereas ML requires human feature engineering as guidance for learning, DL is based on self-learning algorithms that can detect and process simple and complex image features.
Figure 2. Hierarchical structure of AI-techniques. Whereas ML requires human feature engineering as guidance for learning, DL is based on self-learning algorithms that can detect and process simple and complex image features.
Diagnostics 12 00799 g002
Figure 3. Sample radiomics workflow (reprinted with permission from [40], Copyright 2019 Springer Nature).
Figure 3. Sample radiomics workflow (reprinted with permission from [40], Copyright 2019 Springer Nature).
Diagnostics 12 00799 g003
Figure 4. Workflow of standard radiology reporting compared to AI-based methods of radiomic and DL. The entire complexity of deep learning is only schematically shown. There is an abundance of different network architectures or CNN which are beyond the scope of this study. This figure only demonstrates a schematic CNN (reprinted under common creative license 4.0 from [44], Copyright 2021 Springer Nature).
Figure 4. Workflow of standard radiology reporting compared to AI-based methods of radiomic and DL. The entire complexity of deep learning is only schematically shown. There is an abundance of different network architectures or CNN which are beyond the scope of this study. This figure only demonstrates a schematic CNN (reprinted under common creative license 4.0 from [44], Copyright 2021 Springer Nature).
Diagnostics 12 00799 g004
Figure 5. Literature selection work-flow. ML–machine-learning. DL–deep learning. up–uniarametric. bp–biparametric. mp–multiparametric.
Figure 5. Literature selection work-flow. ML–machine-learning. DL–deep learning. up–uniarametric. bp–biparametric. mp–multiparametric.
Diagnostics 12 00799 g005
Figure 6. “Examples of lesion detection. The left two columns show the input T2WI and ADC map, respectively. The right two columns show the FocalNet-predicted lesion probability map and detection points (green crosses) with reference lesion annotation (red contours), respectively. (a) Patient at age 66, with a prostate cancer (PCa) lesion at left anterior peripheral zone with Gleason Group 5 (Gleason Score 4 + 5). (b) Patient at age 68, with a PCa lesion at left posterolateral peripheral zone with Gleason Group 2 (Gleason Score 3 + 4). (c) Patient at age 69, with a PCa lesion at right posterolateral peripheral zone with Gleason Group 3 (Gleason Score 4 + 3). ADC = apparent diffusion coefficient; T2WI = T2-weighted imaging“(reprinted with permission from [72], Copyright 2021 John Wiley and Sons).
Figure 6. “Examples of lesion detection. The left two columns show the input T2WI and ADC map, respectively. The right two columns show the FocalNet-predicted lesion probability map and detection points (green crosses) with reference lesion annotation (red contours), respectively. (a) Patient at age 66, with a prostate cancer (PCa) lesion at left anterior peripheral zone with Gleason Group 5 (Gleason Score 4 + 5). (b) Patient at age 68, with a PCa lesion at left posterolateral peripheral zone with Gleason Group 2 (Gleason Score 3 + 4). (c) Patient at age 69, with a PCa lesion at right posterolateral peripheral zone with Gleason Group 3 (Gleason Score 4 + 3). ADC = apparent diffusion coefficient; T2WI = T2-weighted imaging“(reprinted with permission from [72], Copyright 2021 John Wiley and Sons).
Diagnostics 12 00799 g006
Table 1. List of include studies and relevant key information.
Table 1. List of include studies and relevant key information.
ReferenceYearMLDLField StrengthTargetNumber of PatientsAgeSS/SP/AccuracyAUCSequences Used
Abdollahi H. et al. [40]2019101.5 TGleason score prediction3373 (51–82) 0.739T2, ADC
Wu M. et al. [48]2019103 TTZ PCA detection4468 ± 793.2%/98.4%0.989 (LR)T2, ADC
Varghese B. et al. [49]2019103 TGrading prediction68
53
86%/72%0.71T2, ADC,
Min X. et al. [50]2019103 Tci/csPCA discrimination TZ and PZ280 84.1%/72.7%0.823T2, ADC, b1500
Toivonen J. et al. [51]2019103 TGleason prediction TZ and PZ6265 (45–73) 0.88T2, b0-b2000, T2mapping
Chen T. et al. [52]2019103 TTumor detection
aggressiveness prediction TZ and PZ
182
199
73 (55–90)98.6/99.2%/98.9% (noPCA vs. PCA)
100/98.25 8/99.1% (ci vs. csPCA)
0.999 (noPCA vs. PCA)
0.933 (ciPCA vs. csPCA)
T2, ADC
Xu M. et al. [53]2019103 TTumor detection33171 (46–94) 0.92 (Radiomics)
0.993 (R + clinical data)
T2, ADC, DWI
Zhong X. et al. [54]2019013 Tci/cs PCA discrimination
DL vs. PIRADS exp. radiologists
140 63.6%/80.6%/72.3%
86.4%/48.0%/86.4%
0.726 (DL)
0.711 (PIRADS v2)
Yuan Y. et al. [55]2019013 Tci/cs PCS discrimination (GS > 7)132
112
–/–/86.9% T2 ax and sag, ADC
Xu H. et al. [56]2019013 TDetection of PIRADS ≥ 3 lesions346 –/–/93.0%0.950T2, ADC, high b-value
Schelb P. et al. [57]2019013 TDL and radiologist for lesion (PIRADS ≥ 3 and 4) detection and segmentation250
62
64 (58–71)
64 (60–69)
98/17% Rad, PIRADS ≥ 3
84/48% Rad, PRIADS ≥ 4
99/25%, DL, PIRADS ≥ 3
83/55%, DL, PIRADS ≥ 4
T2, ADC, DWI
Montoya Perez I. et al. [58]2020103 TDetection of csPCA with bpMRI, RNA and clinical data8065 ± 7.1 0.92T2, DWI
Hou Y. et al. [59].2020103 TcsPCA in PIRADS 3 identification in TZ and PZ26366.8 ± 11.4 0.89T2, ADC, b1500
Mehralivand S. et al. [60] 2020103 TDetection csPCA in TZ and PZ236 50.8%/–/– (TZ, MRI)
61.8%/–/– (TZ, DL)
0.749 (MRI)
0.775 (DL)
T2, b1500
Gong L et al. [61]2020103 Tci/cs PCA discrimination 326
163
73.8%/65.8%/69.9%0.788T2, ADC, b800
Bleker J. et al. [62]2020103 Tci/cs PCA discrimination in PZ20666 (48–83) 0.870 (mpMRI)
0.816 (bpMRI)
T2, ADC, DWI, (DCE)
Zong W. et al. [63]2020013 TCNN optimization367 100/92%0.840T2, ADC, b0
Sanford T. et al. [64]2020013 TAutomated PIRADS classification compared to radiologist68767 (46–89) T2, ADC, high b-value
Brunese L. et al. [65]2020111.5 TGleason score prediction52 –/–/98% T2, DCE
Chen Y. et al. [66]2020013 TProstate and cancer segmentation13668 (49–62)75.1/99.9% T2, ADC, b1200
Winkel D.J. et al. [67]2020013 TbpMRI PCA Screening4958 (45–75)87/50% T2, ADC, b2000
Arif M. et al. [68]2020013 TDetection of csPCA in AS29268 (62–72)92/76%0.89T2, ADC, b800
He D. et al. [69]2021103 TTumor detection
Prediction ECE
Prediction PSM
45965 (30–89) 0.863
0.905 (integrated model)
T2, ADC
Vente C. et al. [70]2021013 TcsPCA detection and grading99
63
T2, ADC
Chen J. et al. [71]2021013 TcsPCA detection and grading25 89.6/90.2%/92.1%0.964T2, T1
Cao R. et al. [72]2021013 TPCA detection and grading126
427
62.4 ± 6.4
61.1 ± 7.1
98/17% PIRADS, ≥3
85/58% PIRADS, ≥4
100/17% Unet ≥ 3
83/58% Unet ≥ 4
T2, ADC
Hou Y. et al. [73]2021013 TECE prediction590
150
103
69.2 (42–86)
69.2 (48–83)
70.2 (52–87)
0.857
0.728
T2, ADC, b1500
Yan Y. et al. [74]2021113 TBCR prediction48569.8 0.802 (C-index)T2
Schelb P. et al. [75]2021013 TcsPCA detection and grading28464 (IQR 61–72)98/17% PIRADS, ≥3
85/55% PIRADS, ≥4
99/24% Unet ≥ 3
83/55% Unet ≥ 4
T2, ADC, b1500
Table 2. Display of study results comparing human and AI-based performance.
Table 2. Display of study results comparing human and AI-based performance.
ReferenceYearMLDLMetricHuman RadiologistAI-Approach
Chen T. et al. [52]201910AUC0.8670.999
Schelb P. et al. [57]201901Sensitivity/Specificity98/17% PIRADS ≥ 3
84/48% PRIADS ≥ 4
99/25% PIRADS ≥ 3
83/55% PIRADS ≥ 4
Mehralivand S. et al. [60] 202010AUC
Sensitivity
0.816
89.6%
0.780
87.9%
Sanford T. et al. [64]202001Cancer detection rates53% PIRADS 3
61% PRIADS 4
92% PIRADS 5
57%, PIRADS 3
60%, PIRADS 4
89% PIRADS 5
Cao R. et al. [72]202101Sensitivity/Specificity98/17% PIRADS, ≥3
85/58% PIRADS, ≥4
100/17% PIRADS, ≥3
83/58% PIRADS, ≥4
Schelb P. et al. [75]202101Sensitivity/Specificity98/17% PIRADS, ≥3
85/55% PIRADS, ≥4
99/24% PIRADS, ≥3
83/55% PIRADS, ≥4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Michaely, H.J.; Aringhieri, G.; Cioni, D.; Neri, E. Current Value of Biparametric Prostate MRI with Machine-Learning or Deep-Learning in the Detection, Grading, and Characterization of Prostate Cancer: A Systematic Review. Diagnostics 2022, 12, 799. https://doi.org/10.3390/diagnostics12040799

AMA Style

Michaely HJ, Aringhieri G, Cioni D, Neri E. Current Value of Biparametric Prostate MRI with Machine-Learning or Deep-Learning in the Detection, Grading, and Characterization of Prostate Cancer: A Systematic Review. Diagnostics. 2022; 12(4):799. https://doi.org/10.3390/diagnostics12040799

Chicago/Turabian Style

Michaely, Henrik J., Giacomo Aringhieri, Dania Cioni, and Emanuele Neri. 2022. "Current Value of Biparametric Prostate MRI with Machine-Learning or Deep-Learning in the Detection, Grading, and Characterization of Prostate Cancer: A Systematic Review" Diagnostics 12, no. 4: 799. https://doi.org/10.3390/diagnostics12040799

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop