Next Article in Journal
Accuracy of Modified Blue-Dye Testing in Predicting Dysphagia in Tracheotomized Critically Ill Patients
Next Article in Special Issue
Evaluation of the Reliability and the Performance of Magnetic Resonance Imaging Radiomics in the Presence of Randomly Generated Irrelevant Features for Prostate Cancer
Previous Article in Journal
The Link between NAFLD and Metabolic Syndrome
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Cancer Segmentation and PI-RADS Classification on Multiparametric MRI Using MiniSegCaps Network

Department of Diagnostic Radiology, University of Hong Kong, Hong Kong SAR, China
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(4), 615; https://doi.org/10.3390/diagnostics13040615
Submission received: 10 January 2023 / Revised: 27 January 2023 / Accepted: 28 January 2023 / Published: 7 February 2023

Abstract

:
MRI is the primary imaging approach for diagnosing prostate cancer. Prostate Imaging Reporting and Data System (PI-RADS) on multiparametric MRI (mpMRI) provides fundamental MRI interpretation guidelines but suffers from inter-reader variability. Deep learning networks show great promise in automatic lesion segmentation and classification, which help to ease the burden on radiologists and reduce inter-reader variability. In this study, we proposed a novel multi-branch network, MiniSegCaps, for prostate cancer segmentation and PI-RADS classification on mpMRI. MiniSeg branch outputted the segmentation in conjunction with PI-RADS prediction, guided by the attention map from the CapsuleNet. CapsuleNet branch exploited the relative spatial information of prostate cancer to anatomical structures, such as the zonal location of the lesion, which also reduced the sample size requirement in training due to its equivariance properties. In addition, a gated recurrent unit (GRU) is adopted to exploit spatial knowledge across slices, improving through-plane consistency. Based on the clinical reports, we established a prostate mpMRI database from 462 patients paired with radiologically estimated annotations. MiniSegCaps was trained and evaluated with fivefold cross-validation. On 93 testing cases, our model achieved a 0.712 dice coefficient on lesion segmentation, 89.18% accuracy, and 92.52% sensitivity on PI-RADS classification (PI-RADS ≥ 4) in patient-level evaluation, significantly outperforming existing methods. In addition, a graphical user interface (GUI) integrated into the clinical workflow can automatically produce diagnosis reports based on the results from MiniSegCaps.

1. Introduction

Prostate cancer is the second leading cause of cancer death in men, with an estimated 1,414,259 new deaths in 2020 worldwide [1]. Prostate MRI is widely used to diagnose clinically significant prostate cancer in biopsy-naïve patients. It has been shown that prebiopsy MRI is superior to systematic biopsy in prostate cancer detection [2]. Multiparametric magnetic resonance imaging (mpMRI), including T2-weighted, diffusion-weighted, and dynamic contrast-enhanced scans, plays an integral role in enhanced visualization of prostate cancer, improved biopsy targeting, and monitoring proactive signs of disease progression [2]. In addition, MRI-guided biopsy is also increasingly adopted clinically for risk assessment, replacing the conventional transrectal ultrasound-guided biopsy [3].
The demand for prostate MRI is growing due to mounting evidence and guideline recommendations, and radiologists are facing a substantial increase in referrals [4]. The prostate imaging reporting and data system (PI-RADS) provides fundamental guidelines for assessing prostate MRI by classifying lesions into risk significance as the score increases [5]. PI-RADSv2 has been shown to help radiologists and specialists detect high-grade prostate lesions with high sensitivity [6]. However, PI-RADS has been hindered by poor inter-reader and intra-reader agreement [7]. Furthermore, PI-RADS reporting requires substantial expert knowledge, and radiologists with less experience had more significant inter-reader variability in PI-RADS scoring [8].
With recent developments in deep learning, convolutional neural networks (CNNs) exceeded human performance in natural image analysis [9], especially in image classification [9] and segmentation [10]. Deep learning-based artificial intelligence (AI) algorithms have shown great promise in medical image segmentation, detection, and classification [11,12,13] in recent years.
Deep learning-based lesion detection and PI-RADS classification algorithms are needed for prostate MRI reporting in clinical practice. Recent studies have shown the feasibility of detecting prostate cancer on mpMRI, using deep neural networks [14,15,16,17]. Some networks distinguished prostate cancer from normal tissues and provided the likelihood of prostate cancer [18,19,20]. Some studies compared the performance of network predictions with radiologists’ PI-RADS reports [18,21]. However, the existing method for PI-RADS classification is semi-automated, as manual lesion contours must be inputted into the model [22]. Additionally, these CNNs require a large amount of annotated data and data augmentation, alleviating class imbalance. Few of them addressed lesion detection and classification tasks jointly in one network [23,24] taking spatial relation between prostate lesion and zonal area into consideration which might be important for PI-RADS classification. The recent study only achieved reasonable performance at a PI-RADS cutoff value ≥ 4 [24] with three independent networks for lesion detection and classification which was time-consuming and required more data to train multiple networks. Therefore, this study aimed to solve above-mentioned issues and develop a multi-task network to segment and predict PI-RADS category efficiently by exploiting spatial relation between the prostate lesion and zonal area using limited data.
Capsule Network (CapsNet) [25] helps mitigate data starvation in deep learning-based medical image analysis due to its promising equivariance properties, representing the spatial/viewpoint variability of an object in a capsule (i.e., vector) format [26]. The basic idea of CapsNet is to encode the part-whole relationships (e.g., scale, locations, orientations, brightnesses) between various entities, i.e., objects, and parts of an object, to achieve viewpoint equivariance [25]. Unlike CNN models that learn all partial features of an object, CapsNet understands the relationship between these features through weights that are dynamically computed in each forward pass [25]. This optimization mechanism, dynamic routing, allows the contribution of parts to the whole object to be weighted differently during training and inference [25]. CapsNet has been applied to medical image segmentation in recent studies [27,28], demonstrating the prominence of hybrid architecture between Capsule-based and traditional neural networks in medical image analysis [27].
For PI-RADS, the categorization depends on not only the dimension, edge morphology, and signal intensity, but also the positional relations of the lesion (e.g., extraprostatic extension/invasion) and its zonal location relative to the transition zone (TZ) and peripheral zone (PZ) [5]. Specifically, each lesion can be scored 1–5 on diffusion-weighted imaging (DWI) and T2-weighted (T2W) MRI, as well as by the presence or absence of dynamic contrast enhancement. The contribution of these scores to the overall PI-RADS assessment differs depending on whether the lesion is located in the TZ or PZ of the prostate. For the TZ lesion, the PI-RADS assessment is primarily determined by the T2W score and sometimes modified by the DWI score. For the PZ lesion, the PI-RADS assessment is determined mainly by the DWI score and modulated by the presence of dynamic contrast enhancement [5]. Such spatial relationships and other lesion properties (location, scale, dimension, etc.) can be encoded and represented by CapsNet in a single capsule vector, which is helpful for prostate cancer detection and classification.
Due to our relatively small database, the model in this study was based on a multi-task network with MiniSeg [29] as the backbone for prostate cancer segmentation and PI-RADS classification. Our model, named MiniSegCaps, is built upon a 2D CapsuleNet block, which considers positional relations between the lesion, TZ, and PZ, and 2D convolutional encoder and decoder. Previous studies for PI-RADS classification based on convolutional blocks mainly used one-hot encoding for multi-class classification [23,24]. Different classes were assumed to be equally distanced, ignoring that cancer is a progressive disease. Inspired by Gleason Score prediction in [29], we adopt the ordinal encoding for different PI-RADS categories to encode the lesion progression into vectors. Ordinal encoded vectors are not mutually orthogonal and can suggest the similarities and differences between PI-RADS categories compared with one-hot encoding [29]. In our work, MiniSegCaps encodes five labels, i.e., four PI-RADS categories plus a normal issue type, into ordinal encoded vectors, predicting the ordinal encoded vector for each pixel using mpMRI. Mini-SegCaps can predict lesion masks, the scope (i.e., ordinal encoding) of PI-RADS categories, benign prostatic hyperplasia (BPH), and lesion malignancy. Inspired by a recurrent fully convolutional network (RFCN) [30], a gated recurrent units (GRU) module [31] leveraging inter-slice spatial dependences is also integrated into MiniSegCaps, to exploit the spatial information across adjacent slices, which are represented by capsule vectors, from the entire volume.
The contributions of our work are summarized as follows. Firstly, we proposed a multi-task network to segment and classify prostate cancer on mpMRI jointly. MiniSegCaps inherits the merits of both the 2D CapsuleNet to exploit spatial information and the 2D convolutional blocks to learn better visual representation. Secondly, for MiniSegCaps, we adopted the ordinal encoding to characterize PI-RADS score and the GRU on capsules to exploit spatial knowledge across slices in the entire volume.

2. Materials and Methods

2.1. MRI Protocol

Approval from the institutional review board was obtained for this retrospective study, including 462 patients who underwent prebiopsy MRI and prostate biopsy for network training and evaluation. Patients with prior radiotherapy or hormonal therapy were excluded. All MR imaging was performed on 3 T and 1.5 T scanners (286 patients on 3 T, 157 patients on 1.5 T; both 3 T and 1.5 T MRI systems were from GE Healthcare) with the standardized clinical mpMRI protocol, including T2 weighted (T2w) and apparent diffusion coefficients (ADC). We used axial T2w turbo spin-echo (TSE) imaging and maps of the ADC using echo-planar imaging (EPI) DWI sequence. For T2w, the echo time (TE)/repetition time (TR) was 80/5400 ms on 3 T, and TE/TR = 105/7100 ms on 1.5 T, respectively. T2w on both 3 T and 1.5 T had parameters: FOV = 348 mm × 348 mm, matrix = 512 × 512, number slice = 40, in-plane resolution = 0.68 mm/pixel, and through-plane resolution = 5 mm/slice. For reduced-FOV DWI, TE/TR = 80/5900 ms on 3 T, and TE/TR = 84/7000 ms on 1.5 T. Other parameters were identical on the two scanners: FOV = 300 mm × 300 mm, matrix size = 256 × 256, in-plane resolution = 1.17 mm/pixel, slice thickness = 4 mm, and three b-values = 0/700/1400 s/mm 2 . The ADC maps were obtained using a linear least-squares curve fitting of pixels (in log scale) in the three b0 and diffusion-weighted images against their corresponding b values.
Inclusion criteria for this study were (1) MRI performed between February 2014 and April 2021, (2) men aged from 37 to 90, and (3) patients diagnosed with a consecutive clinical examination or participation in our active surveillance program. We established a cohort of 569 patients with mp-MRI, including T2w, diffusion-weighted, and dynamic contrast material-enhanced MRIs performed at our institution. In this study, we only used two mpMRI sequences including T2w and diffusion-weighted imaging (DWI) due to the limited function of dynamic contrast material-enhanced (DCE) imaging in mpMRIs in the update of PI-RADSv2 [5]. Apparent diffusion coefficient (ADC) maps used in this study were obtained from corresponding DWI data. For this study, we focused on the 494/569 patients with one or more detectable prostate cancer lesions classified by radiologists due to PI-RADS score according to PI-RADSv2 [5]. Exclusion criteria were a history of treatment for prostate cancer (including antihormonal therapy, radiation therapy, focal therapy, and prostatectomy) and incomplete MRI sequences (either missing T2w or ADC images). After excluding 32 patients based on the exclusion criteria, imaging data from 462 patients due to PI-RADS score ≥ 1 were used in this study; each case has both T2w and ADC images.
PI-RADS scores for study cases were obtained from MRI reports based on PI-RADS interpretation of multiparametric MRI performed by four board-certified radiologists during the clinical routine. In addition, one radiologist reviewed all scans and manually delineated the contours of lesions on T2w and ADC images to provide ground truths (GT) based on clinical reports and their accompanying sector map diagrams. Note that completely encapsulated nodules of PI-RADS score = 1 were also labeled and grouped into one category (PI-RADS 1/2) with those of PI-RADS score = 2. Statistical details of the labels in each category were summarized in the table of Section 3.1. As in Figure 1, zonal masks of prostates on T2 images were obtained using pretrained UNet with an average dice of 0.89 on two public datasets (PROSTATEX [32], NCI-ISBI-2013 [33]). Preprocessing operations include resampling, normalization, cropping to the prostate region based on obtained zonal masks, and registration between T2w and ADC images, etc. (Figure 1).

2.2. MiniSegCaps for Prostate Cancer Detection and PI-RADS Classification

In this work, we proposed a multi-task network, MiniSegCaps, which inherits the merits of both CapsNet and CNNs. MiniSegCaps is an end-to-end multi-class CNN to jointly segment prostate lesions and predict their PI-RADS categories. MiniSegCaps adopts MiniSeg [34] as the backbone due to its outstanding performance on a small database, following the Unet-like encoder–decoder architecture [35].
As shown in Figure 2, a multi-task network MiniSegCaps with two predictive branches was proposed for detecting and classifying lesions. A concatenation of T2W, ADC, and zonal mask creates the model’s three-channel inputs. The decoder was one output branch trained for predicting lesion masks and PI-RADS scoping categories under the supervision of ordinal encoded lesion and BPH masks. Another output branch using two capsule layers (caps-branch) was attached to the end of the encoder to produce specific malignancy, i.e., low-grade or high-grade lesion. Reconstructed feature maps from the caps branch were also integrated into the decoder as an attention map to improve the performance of the decoder branch.

2.2.1. Capsule Predictive Branch

The encoder and decoder of MiniSeg [29] extract high-dimensional features from inputs and generate segmentation, respectively. The model takes T2W, ADC, and zonal mask as its three-channel input. It converts the image information to high-dimensional features using a set of convolutional blocks in the encoder that capsules can further process. The learned features from the encoder are reshaped into a grid of H × W × D capsules, each represented as a single 256-dimensional vector. The capsule predictive branch contains two convolutional capsule layers capable of encoding spatial information of objects in the capsule vector. The number of capsule types in the last convolutional capsule layer equals the number of categories in the segmentation, which can be further supervised by a margin loss [25]. This branch is designed to predict the binary high-grade/low-grade PI-RADS categories. Three fully connected (FC) layers followed by the last convolution-capsule layer are also adopted to reconstruct the input features of the capsule predictive branch as in previous work [25]. The reconstructed features are also integrated into the input of the decoder by multiplying with the encoder’s learning components to focus the decoder’s attention on features relevant to a PI-RADS category.

2.2.2. Ordinal Encoding for PI-RADS Categories

We use the decoder of MiniSeg to produce lesion segmentation and PI-RADS scoping categories under the supervision of multi-channel GT masks with ordinal encoding. A conventional multi-class encoder converts each label into a one-hot vector and predicts the one-hot vector through the multi-class output [26]. However, one-hot encoding assumes that different labels are independent of each other, and thus, the cross-entropy loss penalizes misclassifications equally, which ignores the progressiveness of PI-RADS categories. Therefore, we adopt an ordinal encoding [29,36,37] for different PI-RADS categories to encode the progressive lesion commonalities into vectors. Specifically, each bit of an ordinal vector identifies a non-mutually exclusive condition; thus, the k-th bit indicates whether the label is from a category greater than or equal to k (Table 1). In our model, the GTs are encoded into 4-channel masks supervising the model to produce 4-channel segmentation masks. Ordinal encoding characterizes the relationships between different labels; commonalities and differences between labels are represented as shared and distinct bits in an ordinal vector, thus allowing the model to learn the commonalities of all lesions and the distinction between different PI-RADS categories simultaneously [29].

2.2.3. Capsule Gated Recurrent Unit (CapsGRU) for Volumetric Information Integration

To achieve coherent lesion segmentation and prediction of PI-RADS across different slices in one volume, we also introduced a gated recurrent unit (GRU) to exploit spatial dependences across adjacent slices, leveraging inter-slice spatial correlation. The GRU is added to the capsule prediction branch between the encoder and decoder, taking the features learned by the first capsule layer as input. This approach differs from the conventional GRU-Unet method, as our CapsGRU applies to capsules, unlike conventional GRU on feature maps. Here, we denote e s as the output of the first capsule layer where s indicates the slice index, i.e., s ϵ 1 , , S . This output consists of (256 × 6 × 6) feature maps. A recurrent mechanism [30] is introduced to extract global features that capture the spatial changes observed when moving from the base to the apex of the prostate by mapping e s into a new set of features, h s = Φ ( h s 1 , e s ) , where Φ ( ) is a non-linear function, and the size of h s is the same as the size of e s . The feature maps learn by the CapsGRU module and are then flowed to the next capsule predictive layer to produce PI-RADS categories.

2.3. Baseline Methods

Baseline methods in our study include U-Net [35], attention U-Net [38], U-Net++ [39], SegNet [40], MiniSeg [34], FocalNet [29]. To compare with MiniSegCaps, all these models are supervised with ordinal encoding GT segmentations during training. To keep consistent with our network, we adopt the same backbone MiniSeg for FocalNet. For a fair comparison, the training and validation workflows in Figure 2, consisting of image preprocessing, intensity normalization and variation, and image augmentation procedures, are applied equally to all methods.

2.4. Model Training and Evaluation

2.4.1. Loss Functions

Four kinds of loss functions, including dice loss, binary cross-entropy (BCE) loss, margin loss, and mean-squared error (MSE) loss, are adopted in our network. The model was supervised with GT segmentations and PI-RADS categories. The dice loss and BCE loss with the weight of 0.5 are applied at the decoder head with multi-channel GT segmentations. The margin loss is applied to the capsule predictive branch with GT PI-RADS categories. We also use masked MSE loss scaled down by 0.0005 in the total loss, as in previous work [25], for reconstructing the capsule predictive branch.

2.4.2. Implementation Details

The total loss is minimized by stochastic gradient descent (Adam) with an initial learning rate of 0.002, decaying by 0.8 every 20 steps. The network is trained for 300 epochs with a batch size of 256. The MRI images used for patch-based training are cropped to the same size of 100 × 100 to incorporate most prostate regions and further decrease the receptive field of lesion detection. In addition to image normalization, common image augmentations, including image shifting, scaling, and flipping, are also applied during the training. The image augmentations are performed for each batch of training images, not the validation images.

2.4.3. Cross-Validation

We train and validate our model using fivefold cross-validation. Each fold consists of 280 or 281 training cases and 70 or 69 cases for validation. Each case contains 9 to 14 slices, and each fold of training and validation sets has around 3300 and 840 slices, respectively.

2.4.4. Statistical Analysis

To calculate the sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), F1 score, and accuracy in detecting prostate cancer, we defined true or false positives/negatives for an index lesion at the per-patient level. For example, a true positive means that a reader (deep learning method or radiologist) has correctly detected prostate cancer at the same location and categorized the lesion with at least the PI-RADS score category. In addition, the sensitivities and specificities of radiologists were compared to those of the deep learning methods.

2.5. Structure Reporting Graphic User Interface (GUI)

To further aid radiologists in prostate cancer diagnosis in clinical practice, we also designed a graphical user interface (GUI) integrated into the overall workflow to produce diagnosis reports of prostate cancer automatically, which contains the predicted lesion mask, lesion visualization on T2 and ADC, predicted probability of each PI-RADS category, position, and the dimension of each lesion, etc. The main steps of the workflow in GUI include image data importation, zonal segmentation, lesion overlay on mpMRIs, image preprocessing (cropping, normalization, etc.), lesion segmentation, PI-RADS classification, and diagnostic report generation (as shown in Figure 3).
GUI components support image processing toolboxes, picture archiving integration, and integration of deep learning libraries. For clinical AI applications of prostate cancer diagnosis, GUI components included prostate cancer segmentation, PI-RADS classification, and quantification. GUI was deployed with Streamlit software (version 1.2.0, https://streamlit.io/, accessed on 1 May 2022), integrating image processing toolboxes for image preprocessing and postprocessing and training deep learning models for zonal segmentation, lesion segmentation, and PI-RADS classification.
From the MiniSeg branch, the ordinal encoding channel with the the largest number of non-zero pixels was chosen as the final cancer segmentation and PI-RADS prediction. In our generated diagnostic report, the lesion location, i.e., peripheral/transition zone and left/right, was obtained from the intersection between the predicted lesion mask and zonal mask and the relative position of the lesion to the midpoint of the image. The lesion dimensions were computed by the product of the image resolution and the number of lesion pixels.

3. Results

3.1. Study Sample Characteristics

As shown in Table 2, the study sample was randomly divided into a subset used for training and a separate test set. In the training sample, 365 men had 3196 lesions, including 441 of 3196 (14%) PI-RADS category 2, 895 of 3196 (28%) PI-RADS category 3, 1098 of 3196 (34%) PI-RADS category 4, and 798 of 3196 (24%) PI-RADS category 5 lesions. The 93 test patients had 1021 lesions, including 142 of 1021 (14%) PI-RADS category 2, 281 of 1021 (27%) PI-RADS category 3, 353 of 1021 (35%) PI-RADS category 4, and 245 of 1021 (24%) PI-RADS category 5 lesions.

3.2. Prostate Lesion Segmentation

We compare our MiniSegCaps with SOTA baseline segmentation approaches in prostate cancer segmentation evaluated on the testing cohort using the Dice coefficient metric. In addition, a combination version of MiniSeg and CapsuleNet, supervised by ordinal encoding GTs, was also implemented to prove the effectiveness of incorporating Capsule layers into MiniSegCaps. Table 3 and Table 4 show the image-level and patient-level evaluation of these algorithms on prostate cancer segmentation and benign nodule segmentation.
Among baseline methods, 2D U-Net, attention U-Net, and U-Net++ have an averaged Dice coefficient of 51% lower than MiniSeg (65%) in image-level evaluation, as well as similar performance in patient-level evaluation, which indicates that a lightweight model (MiniSeg) performs better in dealing with a small database. Both MiniSegCaps and MiniSegCaps without CapsGRU in Table 3 outperform MiniSeg, SegNet, and FocalNet by a large margin. It indicates that our MiniSegCaps integrating Capsule layers can better identify prostate cancer from normal tissues by learning the relative spatial information of prostate cancer to different anatomical structures. Moreover, our MiniSegCaps obtained better results than MiniSegCaps w/o CapsGRU. As expected, CapsGRU in MiniSegCaps captured the spatial information across adjacent slices, boosting the prostate cancer segmentation performance. Our proposed MiniSegCaps finally improved the performance of prostate cancer segmentation by an average of 5% in the Dice coefficient compared with MiniSeg.
Figure 4 illustrates a visual comparison of the cropped T2, cropped ADC, cropped zonal mask, lesion ground truth (yellow contour), and predicted lesion mask by MiniSegCaps (red contour) on T2 and ADC images among five different cases. Our model successfully produced satisfactory segmentation of prostate cancer and revealed the spatial relationship between the zonal mask, lesion on T2, and ADC, which might help lesion location and classification. We also obtained consistent segmentations across adjacent slices within one volume, as shown in Figure 5. The results indicate that CapsGRU in MiniSegCaps helped capture spatial information and achieve better segmentations across adjacent slices.
Moreover, the results of benign nodule segmentation shown in Table 4 indicate that our model achieved the best performance among these algorithms. Figure 6 also shows a visual comparison of benign nodule ground truth (green contour) and predicted segmentation (blue contour) by deep learning models on T2 and ADC images among different cases.

3.3. PI-RADS Category Prediction

The quantitative results of SOTA baseline methods and our methods in prostate cancer classification in terms of both image-level and patient-level performance in the testing cohort are shown in Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12.
For PI-RADS scoping classification (Table 6, Table 8 and Table 10), the average accuracy of the three categories produced by baseline methods was 57% (PI-RADS ≥ 3), 63% (PI-RADS ≥ 4), 65% (PI-RADS = 5) in the patient-level evaluation, respectively, slightly higher than those results in the image-level evaluation. Due to the consideration of the spatial relationship between objects, both of the last two models integrated capsule layers in Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9 outperform the baseline methods substantially, improving the accuracy of classes (PI-RADS ≥ 3, PI-RADS ≥ 4, PI-RADS = 5) by 15%, 21%, 8%, respectively, and improving the sensitivities of (PI-RADS ≥ 4, PI-RADS = 5) by 5% on average, compared to those of MiniSeg in the patient-level evaluation. Furthermore, our MiniSegCaps consisting of Conv encoder, DeConv decoder with fused feature inputs, and Capsule predictive branch with CapsGRU, perform better in all PI-RADS classes, especially in patient-level evaluation, than MiniSegCaps w/o CapsGRU, the combination version of MiniSeg and CapsuleNet, which contains Conv encoder, DeConv decoder with fused feature inputs and capsule predictive branch. The CapsGRU in MiniSegCaps also improved consistency across adjacent slices to improve the PI-RADS classification results as expected. As a result, our MiniSegCaps achieved the best performance in all PI-RADS scoping classes. MiniSegCaps also improved the accuracy of PI-RADS scoping classification by an average of 15% in patient-level evaluation compared with MiniSeg.
For binary high-grade/low-grade PI-RADS classification (Table 11 and Table 12), the patient-level accuracy and sensitivity of the class (PI-RADS = 4/5) produced by MiniSegCaps were 71.56% and 76.32%, respectively. In addition, the CapsGRU in MiniSegCaps further improved the overall performance of binary high-grade/low-grade lesion differentiation compared with MiniSegCaps w/o CapsGRU.

4. Discussion

In this study, we presented a pipeline for automatic diagnosis of prostate cancer and a novel deep learning method, MiniSegCaps, for prostate cancer segmentation and PI-RADS classification on mpMRI. Our MiniSegCaps, a multi-branch network consisting of capsule predictive layers, exploited the spatial relationships between objects on mpMRI, and a CapsGRU to utilize spatial information across adjacent slices, predicting lesion segmentation and PI-RADS classifications jointly. We trained and validated MiniSegCaps under fivefold cross-validation using 462 pre-operative mpMRIs with annotations of all MRI-visible prostate cancer lesions and benign nodules (e.g., BPH). Experimental results show that our MiniSegCaps outperformed all six CNN-based baseline methods in PI-RADS classification, especially for PI-RADS ≥ 4, which is of great value for diagnosing clinically significant prostate cancer. For lesion segmentation, MiniSegCaps also achieved better performance than baseline methods. We also deployed a structured report GUI integrated into the overall workflow to automatically produce a prostate MRI report, which contained the predicted lesion mask, lesion visualization on T2 and ADC, and predicted probability of each PI-RADS category, position, and the dimension of each lesion, etc.
Previous studies for prostate cancer diagnosis on mpMRI mainly focused on prostate cancer detection based on classic segmentation networks such as U-Net [14,15,16,17,19,20,21]. They showed reasonable performance in distinguishing prostate cancer from normal tissues [14,15,16,17,19,20,21]. However, few considered the relative spatial information of prostate cancer to different anatomical structures to help identify prostate cancer from other anatomical tissues. Furthermore, most of them require a large amount of annotated data which are costly and difficult to acquire in practice. Instead, our MiniSegCaps can integrate anatomical knowledge by exploiting the spatial relationship between objects learned by the capsule branch. It can also be practicable on a small database due to the promising equivariance properties of capsule layers modeling the spatial/viewpoint variability of an object in the image and a lightweight MiniSeg backbone. Furthermore, the CapsGRU in our model is designed to utilize spatial information across adjacent slices, which could reduce flaws in the 2D model when dealing with a series of slices in volume. The experimental results in this study demonstrate that our MiniSegCaps achieved the best performance on prostate cancer detection compared with those classic segmentation networks.
The previous algorithms for PI-RADS classification did not address prostate lesion detection and required manual lesion contouring by radiologists [22]. However, recent work [21,24] comparing prostate cancer detection and PI-RADS classification between deep learning methods and radiologists showed comparable performance between classic CNNs and radiologists. However, the trained models in these studies achieved prostate cancer detection in separate steps. They cannot predict specific lesion segmentations [21], and mainly achieved good performance at a PI-RADS cutoff value ≥ 4. Furthermore, those studies only achieved classification on PI-RADS 3, 4, and 5, not including PI-RADS 1/2. Instead, our multi-branch MiniSegCaps can simultaneously produce prostate cancer masks by the decoder branch and the binary high-grade/low-grade PI-RADS categories (PI-RADS 1/2/3, 4/5) by the Capsule-predictive branch in an end-to-end diagnosis pipeline. Our MiniSegCaps also considered the anatomical priors by learning spatial relationships between objects and achieved the best performance on PI-RADS scoping classification compared with those existing deep neural networks. FocalNet [29] was a multi-class segmentation network for prostate cancer detection and Gleason scores prediction on mpMRIs with focal loss and mutual finding loss. We also applied this method to our prostate cancer segmentation and PI-RADS classification tasks for comparison. The experimental results demonstrate that our MiniSegCaps outperformed the FocalNet in lesion segmentation and PI-RADS classification due to exploiting the spatial relationship between objects for PI-RADS prediction by capsule predictive layers in comparison with FocalNet. A more recent study [23] used 3D cascaded U-Net, residual network architecture for prostate cancer detection, and PI-RADS classification, which required more training data than the 2D networks. Although it achieved acceptable performance in lesion segmentation (0.359 dice coefficient), the overall performance in the PI-RADS classification was unsatisfactory (30.8% accuracy). We tried a residual network (ResNet18 [41]) for PI-RADS classification based on segmented lesions. According to our experiments, the trained model was always prone to over-fitting during the training process, and the corresponding results on the testing cohort were unsatisfactory (30.56% accuracy). Instead, our MiniSegCaps showed prominence in dealing with a small database and achieved better performance in both lesion segmentation (0.712 dice coefficient) and PI-RADS classification (71.56% accuracy) in patient-level evaluation.
Our MiniSegCaps achieved satisfactory prostate cancer detection and PI-RADS classification compared with state-of-the-art methods, especially for cutoff at PI-RADS 3, with clinical importance in disease management. Distinguishing PI-RADS 2 and 3 is challenging for radiologists due to the insignificant difference in hypointense and homogeneous signal intensity on mpMRI. Meanwhile, determining PI-RADS 2 and 3 plays a pivotal role in the differential diagnosis of prostate cancer as it directly affects the clinical decision-making, i.e., biopsy or not. Specifically, prostate cancer with PI-RADS ≥ 3 usually requires a further biopsy to assess lesion aggressiveness by giving a histologically assigned Gleason score (GS) in clinical practice [2]. Therefore, accurate differentiation of PI-RADS 2/3 based on mpMRI reduces unnecessary biopsies [3].
Our multi-task MiniSegCaps can jointly predict prostate cancer masks and PI-RADS categories and achieve satisfactory performance by exploiting spatial relationship between objects. The proposed framework could assist inexperienced readers or non-experts in providing consultations about the prostate cancer contour and PI-RADS categorical probabilities. However, there are also some limitations to this study: (1) We only achieved satisfactory results on PI-RADS scoping and binary high-grade/low-grade classification, not including specific PI-RADS categories due to the sample size limitation; (2) The performance of lesion segmentation and PI-RADS classification on low scores needs further improvement; (3) This study only conducted experiments on local data from one imaging center, not including data from multiple centers. These problems are expected to be solved by incorporating more cases of low PI-RADS categories and enlarging the database from multiple centers. Incorporating the public data from different sites into our database, e.g., those from PROSTATEx [32] and NIH cancer image archive [42], to enlarge the database for PI-RADS classification is currently underway in our lab. In the future, we plan to extend our work to multi-center datasets, further improving the performance of our model on multi-center datasets by incorporating meta learning [43] or few-shot learning [44].

5. Conclusions

Our MiniSegCaps predicted lesion segmentation and PI-RADS classification jointly. As a result, our model achieved the best performance in prostate cancer segmentation and PI-RADS classification compared with state-of-the-art methods, especially for PI-RADS ≥ 3, which was highly important in clinical decision-making.

Author Contributions

Conceptualization, P.C. and V.V.; methodology, W.J. and P.C.; software, W.J. and Y.M.; data curation, Y.L. and V.V.; writing—original draft preparation, W.J.; writing—review and editing, W.J., P.C., V.V., and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Hong Kong Health and Medical Research Fund 07182706.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Acknowledgments

The authors also want to thank Yik-Chung Wu (ycwu@eee.hku.hk) and Yunqi Wang (yunqi9@connect.hku.hk) at Department of Electric and Electronic Engineering, the University of Hong Kong, for the insightful discussion on the CapsuleNet.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ferlay, J.; Ervik, M.; Lam, F.; Colombet, M.; Mery, L.; Piñeros, M.; Znaor, A.; Soerjomataram, I.; Bray, F. Global Cancer Observatory: Cancer Today; International Agency for Research on Cancer: Lyon, France, 2020.
  2. Ahmed, H.U.; Bosaily, A.E.S.; Brown, L.C.; Gabe, R.; Kaplan, R.; Parmar, M.K.; Collaco-Moraes, Y.; Ward, K.; Hindley, R.G.; Freeman, A.; et al. Diagnostic accuracy of multi-parametric MRI and TRUS biopsy in prostate cancer (PROMIS): A paired validating confirmatory study. Lancet 2017, 389, 815–822. [Google Scholar] [CrossRef]
  3. Kasivisvanathan, V.; Rannikko, A.S.; Borghi, M.; Panebianco, V.; Mynderse, L.A.; Vaarala, M.H.; Briganti, A.; Budäus, L.; Hellawell, G.; Hindley, R.G.; et al. MRI-targeted or standard biopsy for prostate-cancer diagnosis. N. Engl. J. Med. 2018, 378, 1767–1777. [Google Scholar] [CrossRef]
  4. Leake, J.L.; Hardman, R.; Ojili, V.; Thompson, I.; Shanbhogue, A.; Hernandez, J.; Barentsz, J. Prostate MRI: Access to and current practice of prostate MRI in the United States. J. Am. Coll. Radiol. 2014, 11, 156–160. [Google Scholar] [CrossRef]
  5. Barrett, T.; Turkbey, B.; Choyke, P.L. PI-RADS version 2: What you need to know. Clin. Radiol. 2015, 70, 1165–1176. [Google Scholar] [CrossRef]
  6. Greer, M.D.; Brown, A.M.; Shih, J.H.; Summers, R.M.; Marko, J.; Law, Y.M.; Sankineni, S.; George, A.K.; Merino, M.J.; Pinto, P.A.; et al. Accuracy and agreement of PIRADSv2 for prostate cancer mpMRI: A multireader study. J. Magn. Reson. Imaging 2017, 45, 579–585. [Google Scholar] [CrossRef]
  7. Smith, C.P.; Harmon, S.A.; Barrett, T.; Bittencourt, L.K.; Law, Y.M.; Shebel, H.; An, J.Y.; Czarniecki, M.; Mehralivand, S.; Coskun, M.; et al. Intra-and interreader reproducibility of PI-RADSv2: A multireader study. J. Magn. Reson. Imaging 2019, 49, 1694–1703. [Google Scholar] [CrossRef]
  8. Sonn, G.A.; Fan, R.E.; Ghanouni, P.; Wang, N.N.; Brooks, J.D.; Loening, A.M.; Daniel, B.L.; To’o, K.J.; Thong, A.E.; Leppert, J.T. Prostate magnetic resonance imaging interpretation varies substantially across radiologists. Eur. Urol. Focus 2019, 5, 592–599. [Google Scholar] [CrossRef]
  9. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 84–90. [Google Scholar] [CrossRef]
  10. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  11. Das, P.K.; Meher, S.; Panda, R.; Abraham, A. An efficient blood-cell segmentation for the detection of hematological disorders. IEEE Trans. Cybern. 2021, 52, 10615–10626. [Google Scholar] [CrossRef]
  12. Das, P.K.; Diya, V.; Meher, S.; Panda, R.; Abraham, A. A systematic review on recent advancements in deep and machine learning based detection and classification of acute lymphoblastic leukemia. IEEE Access 2022, 10, 81741–81763. [Google Scholar] [CrossRef]
  13. Das, P.K.; Meher, S.; Panda, R.; Abraham, A. A review of automated methods for the detection of sickle cell disease. IEEE Rev. Biomed. Eng. 2019, 13, 309–324. [Google Scholar] [CrossRef] [PubMed]
  14. Chen, Y.; Xing, L.; Yu, L.; Bagshaw, H.P.; Buyyounouski, M.K.; Han, B. Automatic intraprostatic lesion segmentation in multiparametric magnetic resonance images with proposed multiple branch UNet. Med. Phys. 2020, 47, 6421–6429. [Google Scholar] [CrossRef] [PubMed]
  15. Saha, A.; Hosseinzadeh, M.; Huisman, H. End-to-end prostate cancer detection in bpMRI via 3D CNNs: Effects of attention mechanisms, clinical priori and decoupled false positive reduction. Med. Image Anal. 2021, 73, 102155. [Google Scholar] [CrossRef] [PubMed]
  16. Saha, A.; Hosseinzadeh, M.; Huisman, H. Encoding clinical priori in 3d convolutional neural networks for prostate cancer detection in bpmri. arXiv 2020, arXiv:2011.00263. [Google Scholar]
  17. Seetharaman, A.; Bhattacharya, I.; Chen, L.C.; Kunder, C.A.; Shao, W.; Soerensen, S.J.; Wang, J.B.; Teslovich, N.C.; Fan, R.E.; Ghanouni, P.; et al. Automated detection of aggressive and indolent prostate cancer on magnetic resonance imaging. Med. Phys. 2021, 48, 2960–2972. [Google Scholar] [CrossRef]
  18. Schelb, P.; Wang, X.; Radtke, J.P.; Wiesenfarth, M.; Kickingereder, P.; Stenzinger, A.; Hohenfellner, M.; Schlemmer, H.P.; Maier-Hein, K.H.; Bonekamp, D. Simulated clinical deployment of fully automatic deep learning for clinical prostate MRI assessment. Eur. Radiol. 2021, 31, 302–313. [Google Scholar] [CrossRef]
  19. Yoo, S.; Gujrathi, I.; Haider, M.A.; Khalvati, F. Prostate cancer detection using deep convolutional neural networks. Sci. Rep. 2019, 9, 1–10. [Google Scholar] [CrossRef]
  20. Sumathipala, Y.; Lay, N.S.; Turkbey, B.; Smith, C.; Choyke, P.L.; Summers, R.M. Prostate cancer detection from multi-institution multiparametric MRIs using deep convolutional neural networks. J. Med. Imaging 2018, 5, 044507. [Google Scholar] [CrossRef]
  21. Schelb, P.; Kohl, S.; Radtke, J.P.; Wiesenfarth, M.; Kickingereder, P.; Bickelhaupt, S.; Kuder, T.A.; Stenzinger, A.; Hohenfellner, M.; Schlemmer, H.P.; et al. Classification of cancer at prostate MRI: Deep learning versus clinical PI-RADS assessment. Radiology 2019, 293, 607–617. [Google Scholar] [CrossRef]
  22. Sanford, T.; Harmon, S.A.; Turkbey, E.B.; Kesani, D.; Tuncer, S.; Madariaga, M.; Yang, C.; Sackett, J.; Mehralivand, S.; Yan, P.; et al. Deep-Learning-Based Artificial Intelligence for PI-RADS Classification to Assist Multiparametric Prostate MRI Interpretation: A Development Study. J. Magn. Reson. Imaging 2020, 52, 1499–1507. [Google Scholar] [CrossRef]
  23. Mehralivand, S.; Yang, D.; Harmon, S.A.; Xu, D.; Xu, Z.; Roth, H.; Masoudi, S.; Sanford, T.H.; Kesani, D.; Lay, N.S.; et al. A Cascaded Deep Learning–Based Artificial Intelligence Algorithm for Automated Lesion Detection and Classification on Biparametric Prostate Magnetic Resonance Imaging. Acad. Radiol. 2022, 29, 1159–1168. [Google Scholar] [CrossRef] [PubMed]
  24. Youn, S.Y.; Choi, M.H.; Kim, D.H.; Lee, Y.J.; Huisman, H.; Johnson, E.; Penzkofer, T.; Shabunin, I.; Winkel, D.J.; Xing, P.; et al. Detection and PI-RADS classification of focal lesions in prostate MRI: Performance comparison between a deep learning-based algorithm (DLA) and radiologists with various levels of experience. Eur. J. Radiol. 2021, 142, 109894. [Google Scholar] [CrossRef]
  25. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic routing between capsules. Adv. Neural Inf. Process. Syst. 2017, 17, 3856–3866. [Google Scholar]
  26. Jiménez-Sánchez, A.; Albarqouni, S.; Mateus, D. Capsule networks against medical imaging data challenges. In Intravascular Imaging and Computer Assisted Stenting and Large-Scale Annotation of Biomedical Data and Expert Label Synthesis; Springer: Berlin/Heidelberg, Germany, 2018; pp. 150–160. [Google Scholar]
  27. Nguyen, T.; Hua, B.S.; Le, N. 3D-UCaps: 3D Capsules Unet for Volumetric Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 548–558. [Google Scholar]
  28. LaLonde, R.; Xu, Z.; Irmakci, I.; Jain, S.; Bagci, U. Capsules for biomedical image segmentation. Med. Image Anal. 2021, 68, 101889. [Google Scholar] [CrossRef] [PubMed]
  29. Cao, R.; Bajgiran, A.M.; Mirak, S.A.; Shakeri, S.; Zhong, X.; Enzmann, D.; Raman, S.; Sung, K. Joint prostate cancer detection and Gleason score prediction in mp-MRI via FocalNet. IEEE Trans. Med. Imaging 2019, 38, 2496–2506. [Google Scholar] [CrossRef]
  30. Poudel, R.P.; Lamata, P.; Montana, G. Recurrent fully convolutional neural networks for multi-slice MRI cardiac segmentation. In Reconstruction, Segmentation, and Analysis of Medical Images; Springer: Berlin/Heidelberg, Germany, 2016; pp. 83–94. [Google Scholar]
  31. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  32. Armato, S.G.; Huisman, H.; Drukker, K.; Hadjiiski, L.; Kirby, J.S.; Petrick, N.; Redmond, G.; Giger, M.L.; Cha, K.; Mamonov, A.; et al. PROSTATEx Challenges for computerized classification of prostate lesions from multiparametric magnetic resonance images. J. Med. Imaging 2018, 5, 044501. [Google Scholar] [CrossRef]
  33. Bloch, N.; Madabhushi, A.; Huisman, H.; Freymann, J.; Kirby, J.; Grauer, M.; Enquobahrie, A.; Jaffe, C.; Clarke, L.; Farahani, K. NCI-ISBI 2013 challenge: Automated segmentation of prostate structures. Cancer Imaging Arch. 2015, 370, 6. [Google Scholar]
  34. Qiu, Y.; Liu, Y.; Li, S.; Xu, J. Miniseg: An extremely minimum network for efficient covid-19 segmentation. arXiv 2020, arXiv:2004.09750. [Google Scholar] [CrossRef]
  35. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  36. Cheng, J.; Wang, Z.; Pollastri, G. A neural network approach to ordinal regression. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–6 June 2008; pp. 1279–1284. [Google Scholar]
  37. Gutiérrez, P.A.; Perez-Ortiz, M.; Sanchez-Monedero, J.; Fernandez-Navarro, F.; Hervas-Martinez, C. Ordinal regression methods: Survey and experimental study. IEEE Trans. Knowl. Data Eng. 2015, 28, 127–146. [Google Scholar] [CrossRef]
  38. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  39. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  40. Badrinarayanan, V.; Handa, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for robust semantic pixel-wise labelling. arXiv 2015, arXiv:1505.07293. [Google Scholar]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  42. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef] [PubMed]
  43. Vanschoren, J. Meta-learning. In Automated Machine Learning: Methods, Systems, Challenges; Springer: Berlin/Heidelberg, Germany, 2019; pp. 35–61. [Google Scholar]
  44. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.; Hospedales, T.M. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar]
Figure 1. The overall pipeline of our work includes four main steps: (1) image preprocessing (registration, normalization, etc.); (2) zonal segmentation and cropping; (3) prostate cancer segmentation and classification; and (4) diagnostic report generation.
Figure 1. The overall pipeline of our work includes four main steps: (1) image preprocessing (registration, normalization, etc.); (2) zonal segmentation and cropping; (3) prostate cancer segmentation and classification; and (4) diagnostic report generation.
Diagnostics 13 00615 g001
Figure 2. Architecture of the proposed MiniSegCaps, consisting of three parts: (1) MiniSeg, a lightweight segmentation network as the backbone for lesion mask prediction; (2) Capsule predictive branch for producing PI-RADS score; (3) CapsGRU module for exploiting spatial information across adjacent slices. The MiniSeg module extracts convolutional features maps from input mpMRIs and produces multi-channel masks for prostate cancer segmentation; learned features (6 × 6 × 256) by the last downsample block of MiniSeg are used as the inputs of capsule predictive branch for PI-RADS classification; with learned capsule feature stacks (8 × 32) by PrimaryCaps as inputs, the CapsGRU module exploits inter-slice spatial information during learning process; reconstructed features (6 × 6 × 256) by three fully connected layers [25] in the Capsule branch are also integrated into the MiniSeg module for better lesion identification.
Figure 2. Architecture of the proposed MiniSegCaps, consisting of three parts: (1) MiniSeg, a lightweight segmentation network as the backbone for lesion mask prediction; (2) Capsule predictive branch for producing PI-RADS score; (3) CapsGRU module for exploiting spatial information across adjacent slices. The MiniSeg module extracts convolutional features maps from input mpMRIs and produces multi-channel masks for prostate cancer segmentation; learned features (6 × 6 × 256) by the last downsample block of MiniSeg are used as the inputs of capsule predictive branch for PI-RADS classification; with learned capsule feature stacks (8 × 32) by PrimaryCaps as inputs, the CapsGRU module exploits inter-slice spatial information during learning process; reconstructed features (6 × 6 × 256) by three fully connected layers [25] in the Capsule branch are also integrated into the MiniSeg module for better lesion identification.
Diagnostics 13 00615 g002
Figure 3. Graphical user interface (GUI) for automatic detection, PI-RADS classification, and diagnostic report generation of prostate cancer on mpMRIs.
Figure 3. Graphical user interface (GUI) for automatic detection, PI-RADS classification, and diagnostic report generation of prostate cancer on mpMRIs.
Diagnostics 13 00615 g003
Figure 4. Visualization of lesion segmentation results among different cases. The yellow contour is the ground truth, and the red contours are from the deep learning predictions.
Figure 4. Visualization of lesion segmentation results among different cases. The yellow contour is the ground truth, and the red contours are from the deep learning predictions.
Diagnostics 13 00615 g004
Figure 5. Visualization of lesion segmentation results on eight slices from one case. The yellow contour is ground truth, and the red contours are predictions from the MiniSegCaps without or with CapsGRU. MiniSegCaps with CapsGRU can better delineate the prostate cancer contours across different slices in one case compared to that without CapsGRU.
Figure 5. Visualization of lesion segmentation results on eight slices from one case. The yellow contour is ground truth, and the red contours are predictions from the MiniSegCaps without or with CapsGRU. MiniSegCaps with CapsGRU can better delineate the prostate cancer contours across different slices in one case compared to that without CapsGRU.
Diagnostics 13 00615 g005
Figure 6. Visualization of benign nodule (e.g., BPH) segmentation results among different cases. The green contour is the ground truth, and the blue contours are from the deep learning predictions.
Figure 6. Visualization of benign nodule (e.g., BPH) segmentation results among different cases. The green contour is the ground truth, and the blue contours are from the deep learning predictions.
Diagnostics 13 00615 g006
Table 1. PI-RADS category encoding for multi-class CNNs.
Table 1. PI-RADS category encoding for multi-class CNNs.
LabelClassOne-Hot EncodingOrdinal Encoding
Non-lesion01 0 0 0 00 0 0 0
PI-RADS 210 1 0 0 01 0 0 0
PI-RADS 320 0 1 0 01 1 0 0
PI-RADS 430 0 0 1 01 1 1 0
PI-RADS 540 0 0 0 11 1 1 1
Table 2. Characteristics and data splitting of the included study samples.
Table 2. Characteristics and data splitting of the included study samples.
Database MRI Scans Total Lesions
+ Benign Nodules
PI-RADS 1/2
(Lesion + Benign Nodules)
PI-RADS 3 PI-RADS 4 PI-RADS 5
Train365626 + 71968 + 719150246162
Test97150 + 26214 + 262375346
In Total462776 + 98182 + 981187299208
Table 3. Comparison of the image-level and patient-level performance between our method and baseline methods in prostate cancer segmentation (PI-RADS ≥ 3).
Table 3. Comparison of the image-level and patient-level performance between our method and baseline methods in prostate cancer segmentation (PI-RADS ≥ 3).
Image LevelPatient Level
ModelDice Coefficient (DSC %)Dice Coefficient (DSC %)
2D U-Net53.3955.20
Attention U-Net48.0149.79
U-Net ++49.5751.06
FocalNet62.5964.38
SegNet53.2056.41
MiniSeg64.8966.73
MiniSegCaps w/o CapsGRU68.2670.89
MiniSegCaps70.1771.16
Table 4. Comparison of the image-level and patient-level performance between our method and baseline methods in benign nodule segmentation (e.g., BPH).
Table 4. Comparison of the image-level and patient-level performance between our method and baseline methods in benign nodule segmentation (e.g., BPH).
Image LevelPatient Level
ModelDice Coefficient (DSC %)Dice Coefficient (DSC %)
2D U-Net47.7049.05
Attention U-Net50.1651.37
U-Net ++46.2648.03
FocalNet55.0156.59
SegNet47.8449.16
MiniSeg55.6356.74
MiniSegCaps w/o CapsGRU56.3557.98
MiniSegCaps57.9659.14
Table 5. Comparison of the image-level performance between our method and baseline methods in prostate cancer classification on PI-RADS ≥ 3.
Table 5. Comparison of the image-level performance between our method and baseline methods in prostate cancer classification on PI-RADS ≥ 3.
Image Level
ModelAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 Score (%) Time per Image (s)
2D U-Net64.1329.2397.0990.4859.2244.190.03
Attention U-Net56.0585.8133.4249.5075.6062.780.03
U-Net ++57.1489.5333.3449.6881.2563.900.03
FocalNet52.3813.0497.2484.3449.5122.580.03
SegNet53.8175.5337.9847.0268.0657.960.03
MiniSeg53.0682.9231.4446.6971.7659.740.03
MiniSegCaps w/o CapsGRU66.3776.2259.0158.1376.8765.960.03
MiniSegCaps67.2077.3759.2059.8976.8767.520.03
Table 6. Comparison of the patient-level performance between our method and baseline methods in prostate cancer classification on PI-RADS ≥ 3.
Table 6. Comparison of the patient-level performance between our method and baseline methods in prostate cancer classification on PI-RADS ≥ 3.
Patient Level
ModelAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 Score (%)Time per Case (s)
2D U-Net65.1430.0497.6392.1360.1345.300.35
Attention U-Net57.2186.5233.9350.9976.0164.170.35
U-Net ++56.0587.6832.7349.0278.2662.880.35
FocalNet51.9812.5296.6280.7249.4121.680.35
SegNet55.7978.5739.2648.4371.6259.920.35
MiniSeg54.6784.6232.5248.1374.0761.350.35
MiniSegCaps w/o CapsGRU68.8079.4460.8460.2879.8268.550.35
MiniSegCaps70.2080.7361.7662.8680.0170.680.35
Table 7. Comparison of the image-level performance between our method and baseline methods in prostate cancer classification on PI-RADS ≥ 4.
Table 7. Comparison of the image-level performance between our method and baseline methods in prostate cancer classification on PI-RADS ≥ 4.
Image Level
ModelAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 Score (%)Time per Image (s)
2D U-Net67.3552.2385.0780.3960.3263.320.03
Attention U-Net57.8588.2932.0652.4176.3665.770.03
U-Net ++58.5471.5451.7143.7877.5654.320.03
FocalNet57.5330.4185.4268.1854.4242.060.03
SegNet63.7080.8244.5562.7765.8570.660.03
MiniSeg67.5485.7152.8859.4682.1170.210.03
MiniSegCaps w/o CapsGRU87.8991.5186.2175.4995.6382.730.03
MiniSegCaps89.1792.7987.5077.4496.3384.430.03
Table 8. Comparison of the patient-level performance between our method and baseline methods in prostate cancer classification on PI-RADS ≥ 4.
Table 8. Comparison of the patient-level performance between our method and baseline methods in prostate cancer classification on PI-RADS ≥ 4.
Patient Level
ModelAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 Score (%)Time per Case (s)
2D U-Net68.3853.1686.4682.3560.8564.620.35
Attention U-Net59.0989.2833.0853.4878.1866.890.35
U-Net ++59.6673.1752.5644.7878.8555.560.35
FocalNet58.2231.3386.6271.2154.4243.520.35
SegNet65.1982.1945.1663.8368.2971.860.35
MiniSeg68.4186.4553.6860.3682.9371.090.35
MiniSegCaps w/o CapsGRU88.0791.6786.4175.8695.6983.010.35
MiniSegCaps89.1892.5287.5878.1696.0684.740.35
Table 9. Comparison of the image-level performance between our method and baseline methods in prostate cancer classification on PI-RADS = 5.
Table 9. Comparison of the image-level performance between our method and baseline methods in prostate cancer classification on PI-RADS = 5.
Image Level
ModelAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 Score (%)Time per Image (s)
2D U-Net56.9329.5382.4260.9855.6939.790.03
Attention U-Net72.8786.8964.1060.2388.6571.140.03
U-Net + + 75.3986.8968.2163.1089.2673.100.03
FocalNet56.9226.3184.6260.7555.9336.720.03
SegNet63.8166.5160.1369.5056.7967.970.03
MiniSeg79.0776.7479.8857.1490.7665.510.03
MiniSegCaps w/o CapsGRU86.9480.4189.1871.8992.9875.910.03
MiniSegCaps87.3481.4489.3672.4893.3376.700.03
Table 10. Comparison of the patient-level performance between our method and baseline methods in prostate cancer classification on PI-RADS = 5.
Table 10. Comparison of the patient-level performance between our method and baseline methods in prostate cancer classification on PI-RADS = 5.
Patient Level
ModelAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 Score (%)Time per Case (s)
2D U-Net63.4469.0542.3735.2671.6751.670.35
Attention U-Net61.1471.4346.5139.4776.9254.520.35
U-Net ++62.1869.7551.7833.9370.0648.130.35
FocalNet58.3338.1074.0753.3360.6144.430.35
SegNet65.7471.1560.7162.7169.3966.670.35
MiniSeg80.7279.1781.3663.3390.5770.370.35
MiniSegCaps w/o CapsGRU87.2581.8289.8679.4191.1880.600.35
MiniSegCaps88.2484.8589.8680.0092.5482.350.35
Table 11. Comparison of the image-level performance between our method and baseline methods in PI-RADS classification of prostate cancer.
Table 11. Comparison of the image-level performance between our method and baseline methods in PI-RADS classification of prostate cancer.
Image Level
ModelPI-RADSAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 Score (%)Time per Image (s)
MiniSegCaps w/o CapsGRU1/2/364.7557.9676.9781.9150.4467.880.03
MiniSegCaps1/2/369.6466.6175.1082.8055.5673.830.03
MiniSegCaps w/o CapsGRU4/564.7576.9757.9650.4481.9160.940.03
MiniSegCaps4/569.6475.1066.6155.5682.8063.870.03
Table 12. Comparison of the patient-level performance between our method and baseline methods in PI-RADS classification of prostate cancer.
Table 12. Comparison of the patient-level performance between our method and baseline methods in PI-RADS classification of prostate cancer.
Patient Level
ModelPI-RADSAccuracy (%)Sensitivity (%)Specificity (%)PPV (%)NPV (%)F1 Score (%) Time per Case (s)
MiniSegCaps w/o CapsGRU1/2/366.3259.6878.7984.0950.9869.810.35
MiniSegCaps1/2/371.5669.0176.3284.4856.8675.970.35
MiniSegCaps w/o CapsGRU4/566.3278.7959.6850.9884.0962.650.35
MiniSegCaps4/571.5676.3269.0156.8684.4865.170.35
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, W.; Lin, Y.; Vardhanabhuti, V.; Ming, Y.; Cao, P. Joint Cancer Segmentation and PI-RADS Classification on Multiparametric MRI Using MiniSegCaps Network. Diagnostics 2023, 13, 615. https://doi.org/10.3390/diagnostics13040615

AMA Style

Jiang W, Lin Y, Vardhanabhuti V, Ming Y, Cao P. Joint Cancer Segmentation and PI-RADS Classification on Multiparametric MRI Using MiniSegCaps Network. Diagnostics. 2023; 13(4):615. https://doi.org/10.3390/diagnostics13040615

Chicago/Turabian Style

Jiang, Wenting, Yingying Lin, Varut Vardhanabhuti, Yanzhen Ming, and Peng Cao. 2023. "Joint Cancer Segmentation and PI-RADS Classification on Multiparametric MRI Using MiniSegCaps Network" Diagnostics 13, no. 4: 615. https://doi.org/10.3390/diagnostics13040615

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop